/* * gstvaapidecoder_h264.c - H.264 decoder * * Copyright (C) 2011-2012 Intel Corporation * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public License * as published by the Free Software Foundation; either version 2.1 * of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free * Software Foundation, Inc., 51 Franklin Street, Fifth Floor, * Boston, MA 02110-1301 USA */ /** * SECTION:gstvaapidecoder_h264 * @short_description: H.264 decoder */ #include "sysdeps.h" #include #include #include #include "gstvaapidecoder_h264.h" #include "gstvaapidecoder_objects.h" #include "gstvaapidecoder_priv.h" #include "gstvaapidisplay_priv.h" #include "gstvaapiobject_priv.h" #define DEBUG 1 #include "gstvaapidebug.h" /* Defined to 1 if strict ordering of DPB is needed. Only useful for debug */ #define USE_STRICT_DPB_ORDERING 0 typedef struct _GstVaapiFrameStore GstVaapiFrameStore; typedef struct _GstVaapiFrameStoreClass GstVaapiFrameStoreClass; typedef struct _GstVaapiDecoderUnitH264 GstVaapiDecoderUnitH264; typedef struct _GstVaapiPictureH264 GstVaapiPictureH264; // Used for field_poc[] #define TOP_FIELD 0 #define BOTTOM_FIELD 1 /* ------------------------------------------------------------------------- */ /* --- H.264 Decoder Units --- */ /* ------------------------------------------------------------------------- */ struct _GstVaapiDecoderUnitH264 { GstVaapiDecoderUnit base; GstH264NalUnit nalu; union { GstH264SPS sps; GstH264PPS pps; GstH264SliceHdr slice_hdr; } data; }; static GstVaapiDecoderUnitH264 * gst_vaapi_decoder_unit_h264_new(guint size) { GstVaapiDecoderUnitH264 *unit; static const GstVaapiMiniObjectClass GstVaapiDecoderUnitH264Class = { sizeof(GstVaapiDecoderUnitH264), (GDestroyNotify)gst_vaapi_decoder_unit_finalize }; unit = (GstVaapiDecoderUnitH264 *) gst_vaapi_mini_object_new(&GstVaapiDecoderUnitH264Class); if (!unit) return NULL; gst_vaapi_decoder_unit_init(&unit->base, size); return unit; } /* ------------------------------------------------------------------------- */ /* --- H.264 Pictures --- */ /* ------------------------------------------------------------------------- */ /* * Extended picture flags: * * @GST_VAAPI_PICTURE_FLAG_IDR: flag that specifies an IDR picture * @GST_VAAPI_PICTURE_FLAG_SHORT_TERM_REFERENCE: flag that specifies * "used for short-term reference" * @GST_VAAPI_PICTURE_FLAG_LONG_TERM_REFERENCE: flag that specifies * "used for long-term reference" * @GST_VAAPI_PICTURE_FLAGS_REFERENCE: mask covering any kind of * reference picture (short-term reference or long-term reference) */ enum { GST_VAAPI_PICTURE_FLAG_IDR = (GST_VAAPI_PICTURE_FLAG_LAST << 0), GST_VAAPI_PICTURE_FLAG_SHORT_TERM_REFERENCE = ( GST_VAAPI_PICTURE_FLAG_REFERENCE), GST_VAAPI_PICTURE_FLAG_LONG_TERM_REFERENCE = ( GST_VAAPI_PICTURE_FLAG_REFERENCE | (GST_VAAPI_PICTURE_FLAG_LAST << 1)), GST_VAAPI_PICTURE_FLAGS_REFERENCE = ( GST_VAAPI_PICTURE_FLAG_SHORT_TERM_REFERENCE | GST_VAAPI_PICTURE_FLAG_LONG_TERM_REFERENCE), }; #define GST_VAAPI_PICTURE_IS_IDR(picture) \ (GST_VAAPI_PICTURE_FLAG_IS_SET(picture, GST_VAAPI_PICTURE_FLAG_IDR)) #define GST_VAAPI_PICTURE_IS_SHORT_TERM_REFERENCE(picture) \ ((GST_VAAPI_PICTURE_FLAGS(picture) & \ GST_VAAPI_PICTURE_FLAGS_REFERENCE) == \ GST_VAAPI_PICTURE_FLAG_SHORT_TERM_REFERENCE) #define GST_VAAPI_PICTURE_IS_LONG_TERM_REFERENCE(picture) \ ((GST_VAAPI_PICTURE_FLAGS(picture) & \ GST_VAAPI_PICTURE_FLAGS_REFERENCE) == \ GST_VAAPI_PICTURE_FLAG_LONG_TERM_REFERENCE) struct _GstVaapiPictureH264 { GstVaapiPicture base; GstH264PPS *pps; GstH264SliceHdr *last_slice_hdr; guint structure; gint32 field_poc[2]; gint32 frame_num; // Original frame_num from slice_header() gint32 frame_num_wrap; // Temporary for ref pic marking: FrameNumWrap gint32 long_term_frame_idx; // Temporary for ref pic marking: LongTermFrameIdx gint32 pic_num; // Temporary for ref pic marking: PicNum gint32 long_term_pic_num; // Temporary for ref pic marking: LongTermPicNum GstVaapiPictureH264 *other_field; // Temporary for ref pic marking: other field in the same frame store guint output_flag : 1; guint output_needed : 1; }; GST_VAAPI_CODEC_DEFINE_TYPE(GstVaapiPictureH264, gst_vaapi_picture_h264); void gst_vaapi_picture_h264_destroy(GstVaapiPictureH264 *picture) { gst_vaapi_picture_destroy(GST_VAAPI_PICTURE(picture)); } gboolean gst_vaapi_picture_h264_create( GstVaapiPictureH264 *picture, const GstVaapiCodecObjectConstructorArgs *args ) { if (!gst_vaapi_picture_create(GST_VAAPI_PICTURE(picture), args)) return FALSE; picture->field_poc[0] = G_MAXINT32; picture->field_poc[1] = G_MAXINT32; picture->output_needed = FALSE; return TRUE; } static inline GstVaapiPictureH264 * gst_vaapi_picture_h264_new(GstVaapiDecoderH264 *decoder) { return (GstVaapiPictureH264 *)gst_vaapi_codec_object_new( &GstVaapiPictureH264Class, GST_VAAPI_CODEC_BASE(decoder), NULL, sizeof(VAPictureParameterBufferH264), NULL, 0, 0); } static inline void gst_vaapi_picture_h264_set_reference( GstVaapiPictureH264 *picture, guint reference_flags, gboolean other_field ) { if (!picture) return; GST_VAAPI_PICTURE_FLAG_UNSET(picture, GST_VAAPI_PICTURE_FLAGS_REFERENCE); GST_VAAPI_PICTURE_FLAG_SET(picture, reference_flags); if (!other_field || !(picture = picture->other_field)) return; GST_VAAPI_PICTURE_FLAG_UNSET(picture, GST_VAAPI_PICTURE_FLAGS_REFERENCE); GST_VAAPI_PICTURE_FLAG_SET(picture, reference_flags); } static inline GstVaapiPictureH264 * gst_vaapi_picture_h264_new_field(GstVaapiPictureH264 *picture) { g_return_val_if_fail(picture, NULL); return (GstVaapiPictureH264 *)gst_vaapi_picture_new_field(&picture->base); } /* ------------------------------------------------------------------------- */ /* --- Frame Buffers (DPB) --- */ /* ------------------------------------------------------------------------- */ struct _GstVaapiFrameStore { /*< private >*/ GstVaapiMiniObject parent_instance; guint structure; GstVaapiPictureH264 *buffers[2]; guint num_buffers; guint output_needed; }; static void gst_vaapi_frame_store_finalize(gpointer object) { GstVaapiFrameStore * const fs = object; guint i; for (i = 0; i < fs->num_buffers; i++) gst_vaapi_picture_replace(&fs->buffers[i], NULL); } static GstVaapiFrameStore * gst_vaapi_frame_store_new(GstVaapiPictureH264 *picture) { GstVaapiFrameStore *fs; static const GstVaapiMiniObjectClass GstVaapiFrameStoreClass = { sizeof(GstVaapiFrameStore), gst_vaapi_frame_store_finalize }; fs = (GstVaapiFrameStore *) gst_vaapi_mini_object_new(&GstVaapiFrameStoreClass); if (!fs) return NULL; fs->structure = picture->structure; fs->buffers[0] = gst_vaapi_picture_ref(picture); fs->buffers[1] = NULL; fs->num_buffers = 1; fs->output_needed = picture->output_needed; return fs; } static gboolean gst_vaapi_frame_store_add(GstVaapiFrameStore *fs, GstVaapiPictureH264 *picture) { guint field; g_return_val_if_fail(fs->num_buffers == 1, FALSE); g_return_val_if_fail(!GST_VAAPI_PICTURE_IS_FRAME(picture), FALSE); g_return_val_if_fail(!GST_VAAPI_PICTURE_IS_FIRST_FIELD(picture), FALSE); gst_vaapi_picture_replace(&fs->buffers[fs->num_buffers++], picture); if (picture->output_flag) { picture->output_needed = TRUE; fs->output_needed++; } fs->structure = GST_VAAPI_PICTURE_STRUCTURE_FRAME; field = picture->structure == GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD ? TOP_FIELD : BOTTOM_FIELD; g_return_val_if_fail(fs->buffers[0]->field_poc[field] == G_MAXINT32, FALSE); fs->buffers[0]->field_poc[field] = picture->field_poc[field]; g_return_val_if_fail(picture->field_poc[!field] == G_MAXINT32, FALSE); picture->field_poc[!field] = fs->buffers[0]->field_poc[!field]; return TRUE; } static gboolean gst_vaapi_frame_store_split_fields(GstVaapiFrameStore *fs) { GstVaapiPictureH264 * const first_field = fs->buffers[0]; GstVaapiPictureH264 *second_field; g_return_val_if_fail(fs->num_buffers == 1, FALSE); first_field->base.structure = GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD; GST_VAAPI_PICTURE_FLAG_SET(first_field, GST_VAAPI_PICTURE_FLAG_INTERLACED); second_field = gst_vaapi_picture_h264_new_field(first_field); if (!second_field) return FALSE; gst_vaapi_picture_replace(&fs->buffers[fs->num_buffers++], second_field); gst_vaapi_picture_unref(second_field); second_field->frame_num = first_field->frame_num; second_field->field_poc[0] = first_field->field_poc[0]; second_field->field_poc[1] = first_field->field_poc[1]; second_field->output_flag = first_field->output_flag; if (second_field->output_flag) { second_field->output_needed = TRUE; fs->output_needed++; } return TRUE; } static inline gboolean gst_vaapi_frame_store_has_frame(GstVaapiFrameStore *fs) { return fs->structure == GST_VAAPI_PICTURE_STRUCTURE_FRAME; } static inline gboolean gst_vaapi_frame_store_has_reference(GstVaapiFrameStore *fs) { guint i; for (i = 0; i < fs->num_buffers; i++) { if (GST_VAAPI_PICTURE_IS_REFERENCE(fs->buffers[i])) return TRUE; } return FALSE; } #define gst_vaapi_frame_store_ref(fs) \ gst_vaapi_mini_object_ref(GST_VAAPI_MINI_OBJECT(fs)) #define gst_vaapi_frame_store_unref(fs) \ gst_vaapi_mini_object_unref(GST_VAAPI_MINI_OBJECT(fs)) #define gst_vaapi_frame_store_replace(old_fs_p, new_fs) \ gst_vaapi_mini_object_replace((GstVaapiMiniObject **)(old_fs_p), \ (GstVaapiMiniObject *)(new_fs)) /* ------------------------------------------------------------------------- */ /* --- H.264 Decoder --- */ /* ------------------------------------------------------------------------- */ G_DEFINE_TYPE(GstVaapiDecoderH264, gst_vaapi_decoder_h264, GST_VAAPI_TYPE_DECODER) #define GST_VAAPI_DECODER_H264_CAST(decoder) \ ((GstVaapiDecoderH264 *)(decoder)) #define GST_VAAPI_DECODER_H264_GET_PRIVATE(obj) \ (G_TYPE_INSTANCE_GET_PRIVATE((obj), \ GST_VAAPI_TYPE_DECODER_H264, \ GstVaapiDecoderH264Private)) struct _GstVaapiDecoderH264Private { GstH264NalParser *parser; GstVaapiPictureH264 *current_picture; GstVaapiDecoderUnitH264 *prev_slice_unit; GstVaapiFrameStore *prev_frame; GstVaapiFrameStore *dpb[16]; guint dpb_count; guint dpb_size; GstVaapiProfile profile; GstVaapiEntrypoint entrypoint; GstVaapiChromaType chroma_type; GstVaapiPictureH264 *short_ref[32]; guint short_ref_count; GstVaapiPictureH264 *long_ref[32]; guint long_ref_count; GstVaapiPictureH264 *RefPicList0[32]; guint RefPicList0_count; GstVaapiPictureH264 *RefPicList1[32]; guint RefPicList1_count; guint nal_length_size; guint mb_width; guint mb_height; gint32 field_poc[2]; // 0:TopFieldOrderCnt / 1:BottomFieldOrderCnt gint32 poc_msb; // PicOrderCntMsb gint32 poc_lsb; // pic_order_cnt_lsb (from slice_header()) gint32 prev_poc_msb; // prevPicOrderCntMsb gint32 prev_poc_lsb; // prevPicOrderCntLsb gint32 frame_num_offset; // FrameNumOffset gint32 frame_num; // frame_num (from slice_header()) gint32 prev_frame_num; // prevFrameNum gboolean prev_pic_has_mmco5; // prevMmco5Pic gboolean prev_pic_structure; // previous picture structure guint is_constructed : 1; guint is_opened : 1; guint is_avcC : 1; guint got_sps : 1; guint got_pps : 1; guint has_context : 1; guint progressive_sequence : 1; }; static gboolean exec_ref_pic_marking(GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture); /* Get number of reference frames to use */ static guint get_max_dec_frame_buffering(GstH264SPS *sps) { guint max_dec_frame_buffering, MaxDpbMbs, PicSizeMbs; /* Table A-1 - Level limits */ switch (sps->level_idc) { case 10: MaxDpbMbs = 396; break; case 11: MaxDpbMbs = 900; break; case 12: MaxDpbMbs = 2376; break; case 13: MaxDpbMbs = 2376; break; case 20: MaxDpbMbs = 2376; break; case 21: MaxDpbMbs = 4752; break; case 22: MaxDpbMbs = 8100; break; case 30: MaxDpbMbs = 8100; break; case 31: MaxDpbMbs = 18000; break; case 32: MaxDpbMbs = 20480; break; case 40: MaxDpbMbs = 32768; break; case 41: MaxDpbMbs = 32768; break; case 42: MaxDpbMbs = 34816; break; case 50: MaxDpbMbs = 110400; break; case 51: MaxDpbMbs = 184320; break; default: g_assert(0 && "unhandled level"); break; } PicSizeMbs = ((sps->pic_width_in_mbs_minus1 + 1) * (sps->pic_height_in_map_units_minus1 + 1) * (sps->frame_mbs_only_flag ? 1 : 2)); max_dec_frame_buffering = MaxDpbMbs / PicSizeMbs; /* VUI parameters */ if (sps->vui_parameters_present_flag) { GstH264VUIParams * const vui_params = &sps->vui_parameters; if (vui_params->bitstream_restriction_flag) max_dec_frame_buffering = vui_params->max_dec_frame_buffering; else { switch (sps->profile_idc) { case 44: // CAVLC 4:4:4 Intra profile case 86: // Scalable High profile case 100: // High profile case 110: // High 10 profile case 122: // High 4:2:2 profile case 244: // High 4:4:4 Predictive profile if (sps->constraint_set3_flag) max_dec_frame_buffering = 0; break; } } } if (max_dec_frame_buffering > 16) max_dec_frame_buffering = 16; else if (max_dec_frame_buffering < sps->num_ref_frames) max_dec_frame_buffering = sps->num_ref_frames; return MAX(1, max_dec_frame_buffering); } static void array_remove_index_fast(void *array, guint *array_length_ptr, guint index) { gpointer * const entries = array; guint num_entries = *array_length_ptr; g_return_if_fail(index < num_entries); if (index != --num_entries) entries[index] = entries[num_entries]; entries[num_entries] = NULL; *array_length_ptr = num_entries; } #if 1 static inline void array_remove_index(void *array, guint *array_length_ptr, guint index) { array_remove_index_fast(array, array_length_ptr, index); } #else static void array_remove_index(void *array, guint *array_length_ptr, guint index) { gpointer * const entries = array; const guint num_entries = *array_length_ptr - 1; guint i; g_return_if_fail(index <= num_entries); for (i = index; i < num_entries; i++) entries[i] = entries[i + 1]; entries[num_entries] = NULL; *array_length_ptr = num_entries; } #endif #define ARRAY_REMOVE_INDEX(array, index) \ array_remove_index(array, &array##_count, index) static void dpb_remove_index(GstVaapiDecoderH264 *decoder, guint index) { GstVaapiDecoderH264Private * const priv = decoder->priv; guint i, num_frames = --priv->dpb_count; if (USE_STRICT_DPB_ORDERING) { for (i = index; i < num_frames; i++) gst_vaapi_frame_store_replace(&priv->dpb[i], priv->dpb[i + 1]); } else if (index != num_frames) gst_vaapi_frame_store_replace(&priv->dpb[index], priv->dpb[num_frames]); gst_vaapi_frame_store_replace(&priv->dpb[num_frames], NULL); } static gboolean dpb_output( GstVaapiDecoderH264 *decoder, GstVaapiFrameStore *fs, GstVaapiPictureH264 *picture ) { picture->output_needed = FALSE; if (fs) { if (--fs->output_needed > 0) return TRUE; picture = fs->buffers[0]; } /* XXX: update cropping rectangle */ return gst_vaapi_picture_output(GST_VAAPI_PICTURE_CAST(picture)); } static inline void dpb_evict(GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, guint i) { GstVaapiFrameStore * const fs = decoder->priv->dpb[i]; if (!fs->output_needed && !gst_vaapi_frame_store_has_reference(fs)) dpb_remove_index(decoder, i); } static gboolean dpb_bump(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPictureH264 *found_picture = NULL; guint i, j, found_index; gboolean success; for (i = 0; i < priv->dpb_count; i++) { GstVaapiFrameStore * const fs = priv->dpb[i]; if (!fs->output_needed) continue; for (j = 0; j < fs->num_buffers; j++) { GstVaapiPictureH264 * const picture = fs->buffers[j]; if (!picture->output_needed) continue; if (!found_picture || found_picture->base.poc > picture->base.poc) found_picture = picture, found_index = i; } } if (!found_picture) return FALSE; success = dpb_output(decoder, priv->dpb[found_index], found_picture); dpb_evict(decoder, found_picture, found_index); return success; } static void dpb_clear(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private * const priv = decoder->priv; guint i; for (i = 0; i < priv->dpb_count; i++) gst_vaapi_frame_store_replace(&priv->dpb[i], NULL); priv->dpb_count = 0; gst_vaapi_frame_store_replace(&priv->prev_frame, NULL); } static void dpb_flush(GstVaapiDecoderH264 *decoder) { while (dpb_bump(decoder)) ; dpb_clear(decoder); } static gboolean dpb_add(GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiFrameStore *fs; guint i, j; // Remove all unused pictures if (!GST_VAAPI_PICTURE_IS_IDR(picture)) { i = 0; while (i < priv->dpb_count) { GstVaapiFrameStore * const fs = priv->dpb[i]; if (!fs->output_needed && !gst_vaapi_frame_store_has_reference(fs)) dpb_remove_index(decoder, i); else i++; } } // Check if picture is the second field and the first field is still in DPB fs = priv->prev_frame; if (fs && !gst_vaapi_frame_store_has_frame(fs)) return gst_vaapi_frame_store_add(fs, picture); // Create new frame store, and split fields if necessary fs = gst_vaapi_frame_store_new(picture); if (!fs) return FALSE; gst_vaapi_frame_store_replace(&priv->prev_frame, fs); gst_vaapi_frame_store_unref(fs); if (!priv->progressive_sequence && gst_vaapi_frame_store_has_frame(fs)) { if (!gst_vaapi_frame_store_split_fields(fs)) return FALSE; } // C.4.5.1 - Storage and marking of a reference decoded picture into the DPB if (GST_VAAPI_PICTURE_IS_REFERENCE(picture)) { while (priv->dpb_count == priv->dpb_size) { if (!dpb_bump(decoder)) return FALSE; } gst_vaapi_frame_store_replace(&priv->dpb[priv->dpb_count++], fs); if (picture->output_flag) { picture->output_needed = TRUE; fs->output_needed++; } } // C.4.5.2 - Storage and marking of a non-reference decoded picture into the DPB else { if (!picture->output_flag) return TRUE; while (priv->dpb_count == priv->dpb_size) { gboolean found_picture = FALSE; for (i = 0; !found_picture && i < priv->dpb_count; i++) { GstVaapiFrameStore * const fs = priv->dpb[i]; if (!fs->output_needed) continue; for (j = 0; !found_picture && j < fs->num_buffers; j++) found_picture = fs->buffers[j]->output_needed && fs->buffers[j]->base.poc < picture->base.poc; } if (!found_picture) return dpb_output(decoder, NULL, picture); if (!dpb_bump(decoder)) return FALSE; } gst_vaapi_frame_store_replace(&priv->dpb[priv->dpb_count++], fs); picture->output_needed = TRUE; fs->output_needed++; } return TRUE; } static inline void dpb_reset(GstVaapiDecoderH264 *decoder, GstH264SPS *sps) { GstVaapiDecoderH264Private * const priv = decoder->priv; priv->dpb_size = get_max_dec_frame_buffering(sps); GST_DEBUG("DPB size %u", priv->dpb_size); } static GstVaapiDecoderStatus get_status(GstH264ParserResult result) { GstVaapiDecoderStatus status; switch (result) { case GST_H264_PARSER_OK: status = GST_VAAPI_DECODER_STATUS_SUCCESS; break; case GST_H264_PARSER_NO_NAL_END: status = GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA; break; case GST_H264_PARSER_ERROR: status = GST_VAAPI_DECODER_STATUS_ERROR_BITSTREAM_PARSER; break; default: status = GST_VAAPI_DECODER_STATUS_ERROR_UNKNOWN; break; } return status; } static void gst_vaapi_decoder_h264_close(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private * const priv = decoder->priv; gst_vaapi_picture_replace(&priv->current_picture, NULL); gst_vaapi_decoder_unit_replace(&priv->prev_slice_unit, NULL); dpb_clear(decoder); if (priv->parser) { gst_h264_nal_parser_free(priv->parser); priv->parser = NULL; } } static gboolean gst_vaapi_decoder_h264_open(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private * const priv = decoder->priv; gst_vaapi_decoder_h264_close(decoder); priv->parser = gst_h264_nal_parser_new(); if (!priv->parser) return FALSE; return TRUE; } static void gst_vaapi_decoder_h264_destroy(GstVaapiDecoderH264 *decoder) { gst_vaapi_decoder_h264_close(decoder); } static gboolean gst_vaapi_decoder_h264_create(GstVaapiDecoderH264 *decoder) { if (!GST_VAAPI_DECODER_CODEC(decoder)) return FALSE; return TRUE; } static guint h264_get_profile(GstH264SPS *sps) { guint profile = 0; switch (sps->profile_idc) { case 66: profile = GST_VAAPI_PROFILE_H264_BASELINE; break; case 77: profile = GST_VAAPI_PROFILE_H264_MAIN; break; case 100: profile = GST_VAAPI_PROFILE_H264_HIGH; break; } return profile; } static guint h264_get_chroma_type(GstH264SPS *sps) { guint chroma_type = 0; switch (sps->chroma_format_idc) { case 1: chroma_type = GST_VAAPI_CHROMA_TYPE_YUV420; break; case 2: chroma_type = GST_VAAPI_CHROMA_TYPE_YUV422; break; case 3: if (!sps->separate_colour_plane_flag) chroma_type = GST_VAAPI_CHROMA_TYPE_YUV444; break; } return chroma_type; } static GstVaapiProfile get_profile(GstVaapiDecoderH264 *decoder, GstH264SPS *sps) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiDisplay * const display = GST_VAAPI_DECODER_DISPLAY(decoder); GstVaapiProfile profile, profiles[2]; guint i, n_profiles = 0; profile = h264_get_profile(sps); if (!profile) return GST_VAAPI_PROFILE_UNKNOWN; profiles[n_profiles++] = profile; switch (profile) { case GST_VAAPI_PROFILE_H264_MAIN: profiles[n_profiles++] = GST_VAAPI_PROFILE_H264_HIGH; break; default: break; } /* If the preferred profile (profiles[0]) matches one that we already found, then just return it now instead of searching for it again */ if (profiles[0] == priv->profile) return priv->profile; for (i = 0; i < n_profiles; i++) { if (gst_vaapi_display_has_decoder(display, profiles[i], priv->entrypoint)) return profiles[i]; } return GST_VAAPI_PROFILE_UNKNOWN; } static GstVaapiDecoderStatus ensure_context(GstVaapiDecoderH264 *decoder, GstH264SPS *sps) { GstVaapiDecoder * const base_decoder = GST_VAAPI_DECODER_CAST(decoder); GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiContextInfo info; GstVaapiProfile profile; GstVaapiChromaType chroma_type; gboolean reset_context = FALSE; guint mb_width, mb_height; profile = get_profile(decoder, sps); if (!profile) { GST_ERROR("unsupported profile_idc %u", sps->profile_idc); return GST_VAAPI_DECODER_STATUS_ERROR_UNSUPPORTED_PROFILE; } if (priv->profile != profile) { GST_DEBUG("profile changed"); reset_context = TRUE; priv->profile = profile; } chroma_type = h264_get_chroma_type(sps); if (!chroma_type || chroma_type != GST_VAAPI_CHROMA_TYPE_YUV420) { GST_ERROR("unsupported chroma_format_idc %u", sps->chroma_format_idc); return GST_VAAPI_DECODER_STATUS_ERROR_UNSUPPORTED_CHROMA_FORMAT; } if (priv->chroma_type != chroma_type) { GST_DEBUG("chroma format changed"); reset_context = TRUE; priv->chroma_type = chroma_type; } mb_width = sps->pic_width_in_mbs_minus1 + 1; mb_height = (sps->pic_height_in_map_units_minus1 + 1) << !sps->frame_mbs_only_flag; if (priv->mb_width != mb_width || priv->mb_height != mb_height) { GST_DEBUG("size changed"); reset_context = TRUE; priv->mb_width = mb_width; priv->mb_height = mb_height; } priv->progressive_sequence = sps->frame_mbs_only_flag; #if 0 /* XXX: we only output complete frames for now */ gst_vaapi_decoder_set_interlaced(base_decoder, !priv->progressive_sequence); #endif gst_vaapi_decoder_set_pixel_aspect_ratio( base_decoder, sps->vui_parameters.par_n, sps->vui_parameters.par_d ); if (!reset_context && priv->has_context) return GST_VAAPI_DECODER_STATUS_SUCCESS; /* XXX: fix surface size when cropping is implemented */ info.profile = priv->profile; info.entrypoint = priv->entrypoint; info.width = sps->width; info.height = sps->height; info.ref_frames = get_max_dec_frame_buffering(sps); if (!gst_vaapi_decoder_ensure_context(GST_VAAPI_DECODER(decoder), &info)) return GST_VAAPI_DECODER_STATUS_ERROR_UNKNOWN; priv->has_context = TRUE; /* Reset DPB */ dpb_reset(decoder, sps); return GST_VAAPI_DECODER_STATUS_SUCCESS; } static void fill_iq_matrix_4x4(VAIQMatrixBufferH264 *iq_matrix, const GstH264PPS *pps) { const guint8 (* const ScalingList4x4)[6][16] = &pps->scaling_lists_4x4; guint i, j; /* There are always 6 4x4 scaling lists */ g_assert(G_N_ELEMENTS(iq_matrix->ScalingList4x4) == 6); g_assert(G_N_ELEMENTS(iq_matrix->ScalingList4x4[0]) == 16); if (sizeof(iq_matrix->ScalingList4x4[0][0]) == 1) memcpy(iq_matrix->ScalingList4x4, *ScalingList4x4, sizeof(iq_matrix->ScalingList4x4)); else { for (i = 0; i < G_N_ELEMENTS(iq_matrix->ScalingList4x4); i++) { for (j = 0; j < G_N_ELEMENTS(iq_matrix->ScalingList4x4[i]); j++) iq_matrix->ScalingList4x4[i][j] = (*ScalingList4x4)[i][j]; } } } static void fill_iq_matrix_8x8(VAIQMatrixBufferH264 *iq_matrix, const GstH264PPS *pps) { const guint8 (* const ScalingList8x8)[6][64] = &pps->scaling_lists_8x8; const GstH264SPS * const sps = pps->sequence; guint i, j, n; /* If chroma_format_idc != 3, there are up to 2 8x8 scaling lists */ if (!pps->transform_8x8_mode_flag) return; g_assert(G_N_ELEMENTS(iq_matrix->ScalingList8x8) >= 2); g_assert(G_N_ELEMENTS(iq_matrix->ScalingList8x8[0]) == 64); if (sizeof(iq_matrix->ScalingList8x8[0][0]) == 1) memcpy(iq_matrix->ScalingList8x8, *ScalingList8x8, sizeof(iq_matrix->ScalingList8x8)); else { n = (sps->chroma_format_idc != 3) ? 2 : 6; for (i = 0; i < n; i++) { for (j = 0; j < G_N_ELEMENTS(iq_matrix->ScalingList8x8[i]); j++) iq_matrix->ScalingList8x8[i][j] = (*ScalingList8x8)[i][j]; } } } static GstVaapiDecoderStatus ensure_quant_matrix(GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture) { GstVaapiPicture * const base_picture = &picture->base; GstH264PPS * const pps = picture->pps; GstH264SPS * const sps = pps->sequence; VAIQMatrixBufferH264 *iq_matrix; base_picture->iq_matrix = GST_VAAPI_IQ_MATRIX_NEW(H264, decoder); if (!base_picture->iq_matrix) { GST_ERROR("failed to allocate IQ matrix"); return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED; } iq_matrix = base_picture->iq_matrix->param; /* XXX: we can only support 4:2:0 or 4:2:2 since ScalingLists8x8[] is not large enough to hold lists for 4:4:4 */ if (sps->chroma_format_idc == 3) return GST_VAAPI_DECODER_STATUS_ERROR_UNSUPPORTED_CHROMA_FORMAT; fill_iq_matrix_4x4(iq_matrix, pps); fill_iq_matrix_8x8(iq_matrix, pps); return GST_VAAPI_DECODER_STATUS_SUCCESS; } static GstVaapiDecoderStatus decode_current_picture(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPictureH264 * const picture = priv->current_picture; if (!picture) return GST_VAAPI_DECODER_STATUS_SUCCESS; if (!exec_ref_pic_marking(decoder, picture)) goto error; if (!dpb_add(decoder, picture)) goto error; if (!gst_vaapi_picture_decode(GST_VAAPI_PICTURE_CAST(picture))) goto error; if (priv->prev_frame && gst_vaapi_frame_store_has_frame(priv->prev_frame)) gst_vaapi_picture_replace(&priv->current_picture, NULL); return GST_VAAPI_DECODER_STATUS_SUCCESS; error: /* XXX: fix for cases where first field failed to be decoded */ gst_vaapi_picture_replace(&priv->current_picture, NULL); return GST_VAAPI_DECODER_STATUS_ERROR_UNKNOWN; } static GstVaapiDecoderStatus parse_sps(GstVaapiDecoderH264 *decoder, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264SPS * const sps = &unit->data.sps; GstH264ParserResult result; GST_DEBUG("parse SPS"); result = gst_h264_parser_parse_sps(priv->parser, &unit->nalu, sps, TRUE); if (result != GST_H264_PARSER_OK) return get_status(result); priv->got_sps = TRUE; return GST_VAAPI_DECODER_STATUS_SUCCESS; } static GstVaapiDecoderStatus parse_pps(GstVaapiDecoderH264 *decoder, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264PPS * const pps = &unit->data.pps; GstH264ParserResult result; GST_DEBUG("parse PPS"); result = gst_h264_parser_parse_pps(priv->parser, &unit->nalu, pps); if (result != GST_H264_PARSER_OK) return get_status(result); priv->got_pps = TRUE; return GST_VAAPI_DECODER_STATUS_SUCCESS; } static GstVaapiDecoderStatus parse_sei(GstVaapiDecoderH264 *decoder, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264SEIMessage sei; GstH264ParserResult result; GST_DEBUG("parse SEI"); memset(&sei, 0, sizeof(sei)); result = gst_h264_parser_parse_sei(priv->parser, &unit->nalu, &sei); if (result != GST_H264_PARSER_OK) { GST_WARNING("failed to parse SEI, payload type:%d", sei.payloadType); return get_status(result); } return GST_VAAPI_DECODER_STATUS_SUCCESS; } static GstVaapiDecoderStatus parse_slice(GstVaapiDecoderH264 *decoder, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264SliceHdr * const slice_hdr = &unit->data.slice_hdr; GstH264ParserResult result; GST_DEBUG("parse slice"); result = gst_h264_parser_parse_slice_hdr(priv->parser, &unit->nalu, slice_hdr, TRUE, TRUE); if (result != GST_H264_PARSER_OK) return get_status(result); return GST_VAAPI_DECODER_STATUS_SUCCESS; } static GstVaapiDecoderStatus decode_sequence_end(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderStatus status; GST_DEBUG("decode sequence-end"); status = decode_current_picture(decoder); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) return status; dpb_flush(decoder); return GST_VAAPI_DECODER_STATUS_SUCCESS; } /* 8.2.1.1 - Decoding process for picture order count type 0 */ static void init_picture_poc_0( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; const gint32 MaxPicOrderCntLsb = 1 << (sps->log2_max_pic_order_cnt_lsb_minus4 + 4); gint32 temp_poc; GST_DEBUG("decode picture order count type 0"); if (GST_VAAPI_PICTURE_IS_IDR(picture)) { priv->prev_poc_msb = 0; priv->prev_poc_lsb = 0; } else if (priv->prev_pic_has_mmco5) { priv->prev_poc_msb = 0; priv->prev_poc_lsb = (priv->prev_pic_structure == GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD ? 0 : priv->field_poc[TOP_FIELD]); } else { priv->prev_poc_msb = priv->poc_msb; priv->prev_poc_lsb = priv->poc_lsb; } // (8-3) priv->poc_lsb = slice_hdr->pic_order_cnt_lsb; if (priv->poc_lsb < priv->prev_poc_lsb && (priv->prev_poc_lsb - priv->poc_lsb) >= (MaxPicOrderCntLsb / 2)) priv->poc_msb = priv->prev_poc_msb + MaxPicOrderCntLsb; else if (priv->poc_lsb > priv->prev_poc_lsb && (priv->poc_lsb - priv->prev_poc_lsb) > (MaxPicOrderCntLsb / 2)) priv->poc_msb = priv->prev_poc_msb - MaxPicOrderCntLsb; else priv->poc_msb = priv->prev_poc_msb; temp_poc = priv->poc_msb + priv->poc_lsb; switch (picture->structure) { case GST_VAAPI_PICTURE_STRUCTURE_FRAME: // (8-4, 8-5) priv->field_poc[TOP_FIELD] = temp_poc; priv->field_poc[BOTTOM_FIELD] = temp_poc + slice_hdr->delta_pic_order_cnt_bottom; break; case GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD: // (8-4) priv->field_poc[TOP_FIELD] = temp_poc; break; case GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD: // (8-5) priv->field_poc[BOTTOM_FIELD] = temp_poc; break; } } /* 8.2.1.2 - Decoding process for picture order count type 1 */ static void init_picture_poc_1( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; const gint32 MaxFrameNum = 1 << (sps->log2_max_frame_num_minus4 + 4); gint32 prev_frame_num_offset, abs_frame_num, expected_poc; guint i; GST_DEBUG("decode picture order count type 1"); if (priv->prev_pic_has_mmco5) prev_frame_num_offset = 0; else prev_frame_num_offset = priv->frame_num_offset; // (8-6) if (GST_VAAPI_PICTURE_IS_IDR(picture)) priv->frame_num_offset = 0; else if (priv->prev_frame_num > priv->frame_num) priv->frame_num_offset = prev_frame_num_offset + MaxFrameNum; else priv->frame_num_offset = prev_frame_num_offset; // (8-7) if (sps->num_ref_frames_in_pic_order_cnt_cycle != 0) abs_frame_num = priv->frame_num_offset + priv->frame_num; else abs_frame_num = 0; if (!GST_VAAPI_PICTURE_IS_REFERENCE(picture) && abs_frame_num > 0) abs_frame_num = abs_frame_num - 1; if (abs_frame_num > 0) { gint32 expected_delta_per_poc_cycle; gint32 poc_cycle_cnt, frame_num_in_poc_cycle; expected_delta_per_poc_cycle = 0; for (i = 0; i < sps->num_ref_frames_in_pic_order_cnt_cycle; i++) expected_delta_per_poc_cycle += sps->offset_for_ref_frame[i]; // (8-8) poc_cycle_cnt = (abs_frame_num - 1) / sps->num_ref_frames_in_pic_order_cnt_cycle; frame_num_in_poc_cycle = (abs_frame_num - 1) % sps->num_ref_frames_in_pic_order_cnt_cycle; // (8-9) expected_poc = poc_cycle_cnt * expected_delta_per_poc_cycle; for (i = 0; i <= frame_num_in_poc_cycle; i++) expected_poc += sps->offset_for_ref_frame[i]; } else expected_poc = 0; if (!GST_VAAPI_PICTURE_IS_REFERENCE(picture)) expected_poc += sps->offset_for_non_ref_pic; // (8-10) switch (picture->structure) { case GST_VAAPI_PICTURE_STRUCTURE_FRAME: priv->field_poc[TOP_FIELD] = expected_poc + slice_hdr->delta_pic_order_cnt[0]; priv->field_poc[BOTTOM_FIELD] = priv->field_poc[TOP_FIELD] + sps->offset_for_top_to_bottom_field + slice_hdr->delta_pic_order_cnt[1]; break; case GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD: priv->field_poc[TOP_FIELD] = expected_poc + slice_hdr->delta_pic_order_cnt[0]; break; case GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD: priv->field_poc[BOTTOM_FIELD] = expected_poc + sps->offset_for_top_to_bottom_field + slice_hdr->delta_pic_order_cnt[0]; break; } } /* 8.2.1.3 - Decoding process for picture order count type 2 */ static void init_picture_poc_2( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; const gint32 MaxFrameNum = 1 << (sps->log2_max_frame_num_minus4 + 4); gint32 prev_frame_num_offset, temp_poc; GST_DEBUG("decode picture order count type 2"); if (priv->prev_pic_has_mmco5) prev_frame_num_offset = 0; else prev_frame_num_offset = priv->frame_num_offset; // (8-11) if (GST_VAAPI_PICTURE_IS_IDR(picture)) priv->frame_num_offset = 0; else if (priv->prev_frame_num > priv->frame_num) priv->frame_num_offset = prev_frame_num_offset + MaxFrameNum; else priv->frame_num_offset = prev_frame_num_offset; // (8-12) if (GST_VAAPI_PICTURE_IS_IDR(picture)) temp_poc = 0; else if (!GST_VAAPI_PICTURE_IS_REFERENCE(picture)) temp_poc = 2 * (priv->frame_num_offset + priv->frame_num) - 1; else temp_poc = 2 * (priv->frame_num_offset + priv->frame_num); // (8-13) if (picture->structure != GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD) priv->field_poc[TOP_FIELD] = temp_poc; if (picture->structure != GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD) priv->field_poc[BOTTOM_FIELD] = temp_poc; } /* 8.2.1 - Decoding process for picture order count */ static void init_picture_poc( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; switch (sps->pic_order_cnt_type) { case 0: init_picture_poc_0(decoder, picture, slice_hdr); break; case 1: init_picture_poc_1(decoder, picture, slice_hdr); break; case 2: init_picture_poc_2(decoder, picture, slice_hdr); break; } if (picture->structure != GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD) picture->field_poc[TOP_FIELD] = priv->field_poc[TOP_FIELD]; if (picture->structure != GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD) picture->field_poc[BOTTOM_FIELD] = priv->field_poc[BOTTOM_FIELD]; picture->base.poc = MIN(picture->field_poc[0], picture->field_poc[1]); } static int compare_picture_pic_num_dec(const void *a, const void *b) { const GstVaapiPictureH264 * const picA = *(GstVaapiPictureH264 **)a; const GstVaapiPictureH264 * const picB = *(GstVaapiPictureH264 **)b; return picB->pic_num - picA->pic_num; } static int compare_picture_long_term_pic_num_inc(const void *a, const void *b) { const GstVaapiPictureH264 * const picA = *(GstVaapiPictureH264 **)a; const GstVaapiPictureH264 * const picB = *(GstVaapiPictureH264 **)b; return picA->long_term_pic_num - picB->long_term_pic_num; } static int compare_picture_poc_dec(const void *a, const void *b) { const GstVaapiPictureH264 * const picA = *(GstVaapiPictureH264 **)a; const GstVaapiPictureH264 * const picB = *(GstVaapiPictureH264 **)b; return picB->base.poc - picA->base.poc; } static int compare_picture_poc_inc(const void *a, const void *b) { const GstVaapiPictureH264 * const picA = *(GstVaapiPictureH264 **)a; const GstVaapiPictureH264 * const picB = *(GstVaapiPictureH264 **)b; return picA->base.poc - picB->base.poc; } static int compare_picture_frame_num_wrap_dec(const void *a, const void *b) { const GstVaapiPictureH264 * const picA = *(GstVaapiPictureH264 **)a; const GstVaapiPictureH264 * const picB = *(GstVaapiPictureH264 **)b; return picB->frame_num_wrap - picA->frame_num_wrap; } static int compare_picture_long_term_frame_idx_inc(const void *a, const void *b) { const GstVaapiPictureH264 * const picA = *(GstVaapiPictureH264 **)a; const GstVaapiPictureH264 * const picB = *(GstVaapiPictureH264 **)b; return picA->long_term_frame_idx - picB->long_term_frame_idx; } /* 8.2.4.1 - Decoding process for picture numbers */ static void init_picture_refs_pic_num( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; const gint32 MaxFrameNum = 1 << (sps->log2_max_frame_num_minus4 + 4); guint i; GST_DEBUG("decode picture numbers"); for (i = 0; i < priv->short_ref_count; i++) { GstVaapiPictureH264 * const pic = priv->short_ref[i]; // (8-27) if (pic->frame_num > priv->frame_num) pic->frame_num_wrap = pic->frame_num - MaxFrameNum; else pic->frame_num_wrap = pic->frame_num; // (8-28, 8-30, 8-31) if (GST_VAAPI_PICTURE_IS_FRAME(picture)) pic->pic_num = pic->frame_num_wrap; else { if (pic->structure == picture->structure) pic->pic_num = 2 * pic->frame_num_wrap + 1; else pic->pic_num = 2 * pic->frame_num_wrap; } } for (i = 0; i < priv->long_ref_count; i++) { GstVaapiPictureH264 * const pic = priv->long_ref[i]; // (8-29, 8-32, 8-33) if (GST_VAAPI_PICTURE_IS_FRAME(picture)) pic->long_term_pic_num = pic->long_term_frame_idx; else { if (pic->structure == picture->structure) pic->long_term_pic_num = 2 * pic->long_term_frame_idx + 1; else pic->long_term_pic_num = 2 * pic->long_term_frame_idx; } } } #define SORT_REF_LIST(list, n, compare_func) \ qsort(list, n, sizeof(*(list)), compare_picture_##compare_func) static void init_picture_refs_fields_1( guint picture_structure, GstVaapiPictureH264 *RefPicList[32], guint *RefPicList_count, GstVaapiPictureH264 *ref_list[32], guint ref_list_count ) { guint i, j, n; i = 0; j = 0; n = *RefPicList_count; do { g_assert(n < 32); for (; i < ref_list_count; i++) { if (ref_list[i]->structure == picture_structure) { RefPicList[n++] = ref_list[i++]; break; } } for (; j < ref_list_count; j++) { if (ref_list[j]->structure != picture_structure) { RefPicList[n++] = ref_list[j++]; break; } } } while (i < ref_list_count || j < ref_list_count); *RefPicList_count = n; } static inline void init_picture_refs_fields( GstVaapiPictureH264 *picture, GstVaapiPictureH264 *RefPicList[32], guint *RefPicList_count, GstVaapiPictureH264 *short_ref[32], guint short_ref_count, GstVaapiPictureH264 *long_ref[32], guint long_ref_count ) { guint n = 0; /* 8.2.4.2.5 - reference picture lists in fields */ init_picture_refs_fields_1(picture->structure, RefPicList, &n, short_ref, short_ref_count); init_picture_refs_fields_1(picture->structure, RefPicList, &n, long_ref, long_ref_count); *RefPicList_count = n; } static void init_picture_refs_p_slice( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPictureH264 **ref_list; guint i; GST_DEBUG("decode reference picture list for P and SP slices"); if (GST_VAAPI_PICTURE_IS_FRAME(picture)) { /* 8.2.4.2.1 - P and SP slices in frames */ if (priv->short_ref_count > 0) { ref_list = priv->RefPicList0; for (i = 0; i < priv->short_ref_count; i++) ref_list[i] = priv->short_ref[i]; SORT_REF_LIST(ref_list, i, pic_num_dec); priv->RefPicList0_count += i; } if (priv->long_ref_count > 0) { ref_list = &priv->RefPicList0[priv->RefPicList0_count]; for (i = 0; i < priv->long_ref_count; i++) ref_list[i] = priv->long_ref[i]; SORT_REF_LIST(ref_list, i, long_term_pic_num_inc); priv->RefPicList0_count += i; } } else { /* 8.2.4.2.2 - P and SP slices in fields */ GstVaapiPictureH264 *short_ref[32]; guint short_ref_count = 0; GstVaapiPictureH264 *long_ref[32]; guint long_ref_count = 0; if (priv->short_ref_count > 0) { for (i = 0; i < priv->short_ref_count; i++) short_ref[i] = priv->short_ref[i]; SORT_REF_LIST(short_ref, i, frame_num_wrap_dec); short_ref_count = i; } if (priv->long_ref_count > 0) { for (i = 0; i < priv->long_ref_count; i++) long_ref[i] = priv->long_ref[i]; SORT_REF_LIST(long_ref, i, long_term_frame_idx_inc); long_ref_count = i; } init_picture_refs_fields( picture, priv->RefPicList0, &priv->RefPicList0_count, short_ref, short_ref_count, long_ref, long_ref_count ); } } static void init_picture_refs_b_slice( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPictureH264 **ref_list; guint i, n; GST_DEBUG("decode reference picture list for B slices"); if (GST_VAAPI_PICTURE_IS_FRAME(picture)) { /* 8.2.4.2.3 - B slices in frames */ /* RefPicList0 */ if (priv->short_ref_count > 0) { // 1. Short-term references ref_list = priv->RefPicList0; for (n = 0, i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->base.poc < picture->base.poc) ref_list[n++] = priv->short_ref[i]; } SORT_REF_LIST(ref_list, n, poc_dec); priv->RefPicList0_count += n; ref_list = &priv->RefPicList0[priv->RefPicList0_count]; for (n = 0, i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->base.poc >= picture->base.poc) ref_list[n++] = priv->short_ref[i]; } SORT_REF_LIST(ref_list, n, poc_inc); priv->RefPicList0_count += n; } if (priv->long_ref_count > 0) { // 2. Long-term references ref_list = &priv->RefPicList0[priv->RefPicList0_count]; for (n = 0, i = 0; i < priv->long_ref_count; i++) ref_list[n++] = priv->long_ref[i]; SORT_REF_LIST(ref_list, n, long_term_pic_num_inc); priv->RefPicList0_count += n; } /* RefPicList1 */ if (priv->short_ref_count > 0) { // 1. Short-term references ref_list = priv->RefPicList1; for (n = 0, i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->base.poc > picture->base.poc) ref_list[n++] = priv->short_ref[i]; } SORT_REF_LIST(ref_list, n, poc_inc); priv->RefPicList1_count += n; ref_list = &priv->RefPicList1[priv->RefPicList1_count]; for (n = 0, i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->base.poc <= picture->base.poc) ref_list[n++] = priv->short_ref[i]; } SORT_REF_LIST(ref_list, n, poc_dec); priv->RefPicList1_count += n; } if (priv->long_ref_count > 0) { // 2. Long-term references ref_list = &priv->RefPicList1[priv->RefPicList1_count]; for (n = 0, i = 0; i < priv->long_ref_count; i++) ref_list[n++] = priv->long_ref[i]; SORT_REF_LIST(ref_list, n, long_term_pic_num_inc); priv->RefPicList1_count += n; } } else { /* 8.2.4.2.4 - B slices in fields */ GstVaapiPictureH264 *short_ref0[32]; guint short_ref0_count = 0; GstVaapiPictureH264 *short_ref1[32]; guint short_ref1_count = 0; GstVaapiPictureH264 *long_ref[32]; guint long_ref_count = 0; /* refFrameList0ShortTerm */ if (priv->short_ref_count > 0) { ref_list = short_ref0; for (n = 0, i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->base.poc <= picture->base.poc) ref_list[n++] = priv->short_ref[i]; } SORT_REF_LIST(ref_list, n, poc_dec); short_ref0_count += n; ref_list = &short_ref0[short_ref0_count]; for (n = 0, i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->base.poc > picture->base.poc) ref_list[n++] = priv->short_ref[i]; } SORT_REF_LIST(ref_list, n, poc_inc); short_ref0_count += n; } /* refFrameList1ShortTerm */ if (priv->short_ref_count > 0) { ref_list = short_ref1; for (n = 0, i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->base.poc > picture->base.poc) ref_list[n++] = priv->short_ref[i]; } SORT_REF_LIST(ref_list, n, poc_inc); short_ref1_count += n; ref_list = &short_ref1[short_ref1_count]; for (n = 0, i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->base.poc <= picture->base.poc) ref_list[n++] = priv->short_ref[i]; } SORT_REF_LIST(ref_list, n, poc_dec); short_ref1_count += n; } /* refFrameListLongTerm */ if (priv->long_ref_count > 0) { for (i = 0; i < priv->long_ref_count; i++) long_ref[i] = priv->long_ref[i]; SORT_REF_LIST(long_ref, i, long_term_frame_idx_inc); long_ref_count = i; } init_picture_refs_fields( picture, priv->RefPicList0, &priv->RefPicList0_count, short_ref0, short_ref0_count, long_ref, long_ref_count ); init_picture_refs_fields( picture, priv->RefPicList1, &priv->RefPicList1_count, short_ref1, short_ref1_count, long_ref, long_ref_count ); } /* Check whether RefPicList1 is identical to RefPicList0, then swap if necessary */ if (priv->RefPicList1_count > 1 && priv->RefPicList1_count == priv->RefPicList0_count && memcmp(priv->RefPicList0, priv->RefPicList1, priv->RefPicList0_count * sizeof(priv->RefPicList0[0])) == 0) { GstVaapiPictureH264 * const tmp = priv->RefPicList1[0]; priv->RefPicList1[0] = priv->RefPicList1[1]; priv->RefPicList1[1] = tmp; } } #undef SORT_REF_LIST static gint find_short_term_reference(GstVaapiDecoderH264 *decoder, gint32 pic_num) { GstVaapiDecoderH264Private * const priv = decoder->priv; guint i; for (i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i]->pic_num == pic_num) return i; } GST_ERROR("found no short-term reference picture with PicNum = %d", pic_num); return -1; } static gint find_long_term_reference(GstVaapiDecoderH264 *decoder, gint32 long_term_pic_num) { GstVaapiDecoderH264Private * const priv = decoder->priv; guint i; for (i = 0; i < priv->long_ref_count; i++) { if (priv->long_ref[i]->long_term_pic_num == long_term_pic_num) return i; } GST_ERROR("found no long-term reference picture with LongTermPicNum = %d", long_term_pic_num); return -1; } static void exec_picture_refs_modification_1( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr, guint list ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; GstH264RefPicListModification *ref_pic_list_modification; guint num_ref_pic_list_modifications; GstVaapiPictureH264 **ref_list; guint *ref_list_count_ptr, ref_list_count, ref_list_idx = 0; guint i, j, n, num_refs; gint found_ref_idx; gint32 MaxPicNum, CurrPicNum, picNumPred; GST_DEBUG("modification process of reference picture list %u", list); if (list == 0) { ref_pic_list_modification = slice_hdr->ref_pic_list_modification_l0; num_ref_pic_list_modifications = slice_hdr->n_ref_pic_list_modification_l0; ref_list = priv->RefPicList0; ref_list_count_ptr = &priv->RefPicList0_count; num_refs = slice_hdr->num_ref_idx_l0_active_minus1 + 1; } else { ref_pic_list_modification = slice_hdr->ref_pic_list_modification_l1; num_ref_pic_list_modifications = slice_hdr->n_ref_pic_list_modification_l1; ref_list = priv->RefPicList1; ref_list_count_ptr = &priv->RefPicList1_count; num_refs = slice_hdr->num_ref_idx_l1_active_minus1 + 1; } ref_list_count = *ref_list_count_ptr; if (!GST_VAAPI_PICTURE_IS_FRAME(picture)) { MaxPicNum = 1 << (sps->log2_max_frame_num_minus4 + 5); // 2 * MaxFrameNum CurrPicNum = 2 * slice_hdr->frame_num + 1; // 2 * frame_num + 1 } else { MaxPicNum = 1 << (sps->log2_max_frame_num_minus4 + 4); // MaxFrameNum CurrPicNum = slice_hdr->frame_num; // frame_num } picNumPred = CurrPicNum; for (i = 0; i < num_ref_pic_list_modifications; i++) { GstH264RefPicListModification * const l = &ref_pic_list_modification[i]; if (l->modification_of_pic_nums_idc == 3) break; /* 8.2.4.3.1 - Short-term reference pictures */ if (l->modification_of_pic_nums_idc == 0 || l->modification_of_pic_nums_idc == 1) { gint32 abs_diff_pic_num = l->value.abs_diff_pic_num_minus1 + 1; gint32 picNum, picNumNoWrap; // (8-34) if (l->modification_of_pic_nums_idc == 0) { picNumNoWrap = picNumPred - abs_diff_pic_num; if (picNumNoWrap < 0) picNumNoWrap += MaxPicNum; } // (8-35) else { picNumNoWrap = picNumPred + abs_diff_pic_num; if (picNumNoWrap >= MaxPicNum) picNumNoWrap -= MaxPicNum; } picNumPred = picNumNoWrap; // (8-36) picNum = picNumNoWrap; if (picNum > CurrPicNum) picNum -= MaxPicNum; // (8-37) for (j = num_refs; j > ref_list_idx; j--) ref_list[j] = ref_list[j - 1]; found_ref_idx = find_short_term_reference(decoder, picNum); ref_list[ref_list_idx++] = found_ref_idx >= 0 ? priv->short_ref[found_ref_idx] : NULL; n = ref_list_idx; for (j = ref_list_idx; j <= num_refs; j++) { gint32 PicNumF; if (!ref_list[j]) continue; PicNumF = GST_VAAPI_PICTURE_IS_SHORT_TERM_REFERENCE(ref_list[j]) ? ref_list[j]->pic_num : MaxPicNum; if (PicNumF != picNum) ref_list[n++] = ref_list[j]; } } /* 8.2.4.3.2 - Long-term reference pictures */ else { for (j = num_refs; j > ref_list_idx; j--) ref_list[j] = ref_list[j - 1]; found_ref_idx = find_long_term_reference(decoder, l->value.long_term_pic_num); ref_list[ref_list_idx++] = found_ref_idx >= 0 ? priv->long_ref[found_ref_idx] : NULL; n = ref_list_idx; for (j = ref_list_idx; j <= num_refs; j++) { gint32 LongTermPicNumF; if (!ref_list[j]) continue; LongTermPicNumF = GST_VAAPI_PICTURE_IS_LONG_TERM_REFERENCE(ref_list[j]) ? ref_list[j]->long_term_pic_num : INT_MAX; if (LongTermPicNumF != l->value.long_term_pic_num) ref_list[n++] = ref_list[j]; } } } #if DEBUG for (i = 0; i < num_refs; i++) if (!ref_list[i]) GST_ERROR("list %u entry %u is empty", list, i); #endif *ref_list_count_ptr = num_refs; } /* 8.2.4.3 - Modification process for reference picture lists */ static void exec_picture_refs_modification( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GST_DEBUG("execute ref_pic_list_modification()"); /* RefPicList0 */ if (!GST_H264_IS_I_SLICE(slice_hdr) && !GST_H264_IS_SI_SLICE(slice_hdr) && slice_hdr->ref_pic_list_modification_flag_l0) exec_picture_refs_modification_1(decoder, picture, slice_hdr, 0); /* RefPicList1 */ if (GST_H264_IS_B_SLICE(slice_hdr) && slice_hdr->ref_pic_list_modification_flag_l1) exec_picture_refs_modification_1(decoder, picture, slice_hdr, 1); } static void init_picture_ref_lists(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private * const priv = decoder->priv; guint i, j, short_ref_count, long_ref_count; short_ref_count = 0; long_ref_count = 0; if (GST_VAAPI_PICTURE_IS_FRAME(priv->current_picture)) { for (i = 0; i < priv->dpb_count; i++) { GstVaapiFrameStore * const fs = priv->dpb[i]; GstVaapiPictureH264 *picture; if (!gst_vaapi_frame_store_has_frame(fs)) continue; picture = fs->buffers[0]; if (GST_VAAPI_PICTURE_IS_SHORT_TERM_REFERENCE(picture)) priv->short_ref[short_ref_count++] = picture; else if (GST_VAAPI_PICTURE_IS_LONG_TERM_REFERENCE(picture)) priv->long_ref[long_ref_count++] = picture; picture->structure = GST_VAAPI_PICTURE_STRUCTURE_FRAME; picture->other_field = fs->buffers[1]; } } else { for (i = 0; i < priv->dpb_count; i++) { GstVaapiFrameStore * const fs = priv->dpb[i]; for (j = 0; j < fs->num_buffers; j++) { GstVaapiPictureH264 * const picture = fs->buffers[j]; if (GST_VAAPI_PICTURE_IS_SHORT_TERM_REFERENCE(picture)) priv->short_ref[short_ref_count++] = picture; else if (GST_VAAPI_PICTURE_IS_LONG_TERM_REFERENCE(picture)) priv->long_ref[long_ref_count++] = picture; picture->structure = picture->base.structure; picture->other_field = fs->buffers[j ^ 1]; } } } for (i = short_ref_count; i < priv->short_ref_count; i++) priv->short_ref[i] = NULL; priv->short_ref_count = short_ref_count; for (i = long_ref_count; i < priv->long_ref_count; i++) priv->long_ref[i] = NULL; priv->long_ref_count = long_ref_count; } static void init_picture_refs( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264SliceHdr *slice_hdr ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPicture * const base_picture = &picture->base; guint i, num_refs; init_picture_ref_lists(decoder); init_picture_refs_pic_num(decoder, picture, slice_hdr); priv->RefPicList0_count = 0; priv->RefPicList1_count = 0; switch (base_picture->type) { case GST_VAAPI_PICTURE_TYPE_P: case GST_VAAPI_PICTURE_TYPE_SP: init_picture_refs_p_slice(decoder, picture, slice_hdr); break; case GST_VAAPI_PICTURE_TYPE_B: init_picture_refs_b_slice(decoder, picture, slice_hdr); break; default: break; } exec_picture_refs_modification(decoder, picture, slice_hdr); switch (base_picture->type) { case GST_VAAPI_PICTURE_TYPE_B: num_refs = 1 + slice_hdr->num_ref_idx_l1_active_minus1; for (i = priv->RefPicList1_count; i < num_refs; i++) priv->RefPicList1[i] = NULL; priv->RefPicList1_count = num_refs; // fall-through case GST_VAAPI_PICTURE_TYPE_P: case GST_VAAPI_PICTURE_TYPE_SP: num_refs = 1 + slice_hdr->num_ref_idx_l0_active_minus1; for (i = priv->RefPicList0_count; i < num_refs; i++) priv->RefPicList0[i] = NULL; priv->RefPicList0_count = num_refs; break; default: break; } } static gboolean init_picture( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPicture * const base_picture = &picture->base; GstH264SliceHdr * const slice_hdr = &unit->data.slice_hdr; priv->prev_frame_num = priv->frame_num; priv->frame_num = slice_hdr->frame_num; picture->frame_num = priv->frame_num; picture->frame_num_wrap = priv->frame_num; picture->output_flag = TRUE; /* XXX: conformant to Annex A only */ base_picture->pts = GST_VAAPI_DECODER_CODEC_FRAME(decoder)->pts; /* Reset decoder state for IDR pictures */ if (unit->nalu.type == GST_H264_NAL_SLICE_IDR) { GST_DEBUG(""); GST_VAAPI_PICTURE_FLAG_SET(picture, GST_VAAPI_PICTURE_FLAG_IDR); dpb_flush(decoder); } /* Initialize slice type */ switch (slice_hdr->type % 5) { case GST_H264_P_SLICE: base_picture->type = GST_VAAPI_PICTURE_TYPE_P; break; case GST_H264_B_SLICE: base_picture->type = GST_VAAPI_PICTURE_TYPE_B; break; case GST_H264_I_SLICE: base_picture->type = GST_VAAPI_PICTURE_TYPE_I; break; case GST_H264_SP_SLICE: base_picture->type = GST_VAAPI_PICTURE_TYPE_SP; break; case GST_H264_SI_SLICE: base_picture->type = GST_VAAPI_PICTURE_TYPE_SI; break; } /* Initialize picture structure */ if (!slice_hdr->field_pic_flag) base_picture->structure = GST_VAAPI_PICTURE_STRUCTURE_FRAME; else { GST_VAAPI_PICTURE_FLAG_SET(picture, GST_VAAPI_PICTURE_FLAG_INTERLACED); if (!slice_hdr->bottom_field_flag) base_picture->structure = GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD; else base_picture->structure = GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD; } picture->structure = base_picture->structure; /* Initialize reference flags */ if (unit->nalu.ref_idc) { GstH264DecRefPicMarking * const dec_ref_pic_marking = &slice_hdr->dec_ref_pic_marking; if (GST_VAAPI_PICTURE_IS_IDR(picture) && dec_ref_pic_marking->long_term_reference_flag) GST_VAAPI_PICTURE_FLAG_SET(picture, GST_VAAPI_PICTURE_FLAG_LONG_TERM_REFERENCE); else GST_VAAPI_PICTURE_FLAG_SET(picture, GST_VAAPI_PICTURE_FLAG_SHORT_TERM_REFERENCE); } init_picture_poc(decoder, picture, slice_hdr); init_picture_refs(decoder, picture, slice_hdr); return TRUE; } /* 8.2.5.3 - Sliding window decoded reference picture marking process */ static gboolean exec_ref_pic_marking_sliding_window(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264PPS * const pps = priv->current_picture->pps; GstH264SPS * const sps = pps->sequence; GstVaapiPictureH264 *ref_picture; guint i, m, max_num_ref_frames; GST_DEBUG("reference picture marking process (sliding window)"); if (!GST_VAAPI_PICTURE_IS_FIRST_FIELD(priv->current_picture)) return TRUE; max_num_ref_frames = sps->num_ref_frames; if (max_num_ref_frames == 0) max_num_ref_frames = 1; if (!GST_VAAPI_PICTURE_IS_FRAME(priv->current_picture)) max_num_ref_frames <<= 1; if (priv->short_ref_count + priv->long_ref_count < max_num_ref_frames) return TRUE; if (priv->short_ref_count < 1) return FALSE; for (m = 0, i = 1; i < priv->short_ref_count; i++) { GstVaapiPictureH264 * const picture = priv->short_ref[i]; if (picture->frame_num_wrap < priv->short_ref[m]->frame_num_wrap) m = i; } ref_picture = priv->short_ref[m]; gst_vaapi_picture_h264_set_reference(ref_picture, 0, TRUE); ARRAY_REMOVE_INDEX(priv->short_ref, m); /* Both fields need to be marked as "unused for reference", so remove the other field from the short_ref[] list as well */ if (!GST_VAAPI_PICTURE_IS_FRAME(priv->current_picture) && ref_picture->other_field) { for (i = 0; i < priv->short_ref_count; i++) { if (priv->short_ref[i] == ref_picture->other_field) { ARRAY_REMOVE_INDEX(priv->short_ref, i); break; } } } return TRUE; } static inline gint32 get_picNumX(GstVaapiPictureH264 *picture, GstH264RefPicMarking *ref_pic_marking) { gint32 pic_num; if (GST_VAAPI_PICTURE_IS_FRAME(picture)) pic_num = picture->frame_num_wrap; else pic_num = 2 * picture->frame_num_wrap + 1; pic_num -= ref_pic_marking->difference_of_pic_nums_minus1 + 1; return pic_num; } /* 8.2.5.4.1. Mark short-term reference picture as "unused for reference" */ static void exec_ref_pic_marking_adaptive_mmco_1( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264RefPicMarking *ref_pic_marking ) { GstVaapiDecoderH264Private * const priv = decoder->priv; gint32 i, picNumX; picNumX = get_picNumX(picture, ref_pic_marking); i = find_short_term_reference(decoder, picNumX); if (i < 0) return; gst_vaapi_picture_h264_set_reference(priv->short_ref[i], 0, GST_VAAPI_PICTURE_IS_FRAME(picture)); ARRAY_REMOVE_INDEX(priv->short_ref, i); } /* 8.2.5.4.2. Mark long-term reference picture as "unused for reference" */ static void exec_ref_pic_marking_adaptive_mmco_2( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264RefPicMarking *ref_pic_marking ) { GstVaapiDecoderH264Private * const priv = decoder->priv; gint32 i; i = find_long_term_reference(decoder, ref_pic_marking->long_term_pic_num); if (i < 0) return; gst_vaapi_picture_h264_set_reference(priv->long_ref[i], 0, GST_VAAPI_PICTURE_IS_FRAME(picture)); ARRAY_REMOVE_INDEX(priv->long_ref, i); } /* 8.2.5.4.3. Assign LongTermFrameIdx to a short-term reference picture */ static void exec_ref_pic_marking_adaptive_mmco_3( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264RefPicMarking *ref_pic_marking ) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPictureH264 *ref_picture; gint32 i, picNumX; for (i = 0; i < priv->long_ref_count; i++) { if (priv->long_ref[i]->long_term_frame_idx == ref_pic_marking->long_term_frame_idx) break; } if (i != priv->long_ref_count) { gst_vaapi_picture_h264_set_reference(priv->long_ref[i], 0, TRUE); ARRAY_REMOVE_INDEX(priv->long_ref, i); } picNumX = get_picNumX(picture, ref_pic_marking); i = find_short_term_reference(decoder, picNumX); if (i < 0) return; ref_picture = priv->short_ref[i]; ARRAY_REMOVE_INDEX(priv->short_ref, i); priv->long_ref[priv->long_ref_count++] = ref_picture; ref_picture->long_term_frame_idx = ref_pic_marking->long_term_frame_idx; gst_vaapi_picture_h264_set_reference(ref_picture, GST_VAAPI_PICTURE_FLAG_LONG_TERM_REFERENCE, GST_VAAPI_PICTURE_IS_FRAME(picture)); } /* 8.2.5.4.4. Mark pictures with LongTermFramIdx > max_long_term_frame_idx * as "unused for reference" */ static void exec_ref_pic_marking_adaptive_mmco_4( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264RefPicMarking *ref_pic_marking ) { GstVaapiDecoderH264Private * const priv = decoder->priv; gint32 i, long_term_frame_idx; long_term_frame_idx = ref_pic_marking->max_long_term_frame_idx_plus1 - 1; for (i = 0; i < priv->long_ref_count; i++) { if (priv->long_ref[i]->long_term_frame_idx <= long_term_frame_idx) continue; gst_vaapi_picture_h264_set_reference(priv->long_ref[i], 0, FALSE); ARRAY_REMOVE_INDEX(priv->long_ref, i); i--; } } /* 8.2.5.4.5. Mark all reference pictures as "unused for reference" */ static void exec_ref_pic_marking_adaptive_mmco_5( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264RefPicMarking *ref_pic_marking ) { GstVaapiDecoderH264Private * const priv = decoder->priv; dpb_flush(decoder); priv->prev_pic_has_mmco5 = TRUE; /* The picture shall be inferred to have had frame_num equal to 0 (7.4.3) */ priv->frame_num = 0; priv->frame_num_offset = 0; picture->frame_num = 0; /* Update TopFieldOrderCnt and BottomFieldOrderCnt (8.2.1) */ if (picture->structure != GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD) picture->field_poc[TOP_FIELD] -= picture->base.poc; if (picture->structure != GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD) picture->field_poc[BOTTOM_FIELD] -= picture->base.poc; picture->base.poc = 0; } /* 8.2.5.4.6. Assign a long-term frame index to the current picture */ static void exec_ref_pic_marking_adaptive_mmco_6( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264RefPicMarking *ref_pic_marking ) { picture->long_term_frame_idx = ref_pic_marking->long_term_frame_idx; gst_vaapi_picture_h264_set_reference(picture, GST_VAAPI_PICTURE_FLAG_LONG_TERM_REFERENCE, FALSE); } /* 8.2.5.4. Adaptive memory control decoded reference picture marking process */ static gboolean exec_ref_pic_marking_adaptive( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264DecRefPicMarking *dec_ref_pic_marking ) { guint i; GST_DEBUG("reference picture marking process (adaptive memory control)"); typedef void (*exec_ref_pic_marking_adaptive_mmco_func)( GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstH264RefPicMarking *ref_pic_marking ); static const exec_ref_pic_marking_adaptive_mmco_func mmco_funcs[] = { NULL, exec_ref_pic_marking_adaptive_mmco_1, exec_ref_pic_marking_adaptive_mmco_2, exec_ref_pic_marking_adaptive_mmco_3, exec_ref_pic_marking_adaptive_mmco_4, exec_ref_pic_marking_adaptive_mmco_5, exec_ref_pic_marking_adaptive_mmco_6, }; for (i = 0; i < dec_ref_pic_marking->n_ref_pic_marking; i++) { GstH264RefPicMarking * const ref_pic_marking = &dec_ref_pic_marking->ref_pic_marking[i]; const guint mmco = ref_pic_marking->memory_management_control_operation; if (mmco < G_N_ELEMENTS(mmco_funcs) && mmco_funcs[mmco]) mmco_funcs[mmco](decoder, picture, ref_pic_marking); else { GST_ERROR("unhandled MMCO %u", mmco); return FALSE; } } return TRUE; } /* 8.2.5 - Execute reference picture marking process */ static gboolean exec_ref_pic_marking(GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture) { GstVaapiDecoderH264Private * const priv = decoder->priv; priv->prev_pic_has_mmco5 = FALSE; priv->prev_pic_structure = picture->structure; if (!GST_VAAPI_PICTURE_IS_REFERENCE(picture)) return TRUE; if (!GST_VAAPI_PICTURE_IS_IDR(picture)) { GstH264DecRefPicMarking * const dec_ref_pic_marking = &picture->last_slice_hdr->dec_ref_pic_marking; if (dec_ref_pic_marking->adaptive_ref_pic_marking_mode_flag) { if (!exec_ref_pic_marking_adaptive(decoder, picture, dec_ref_pic_marking)) return FALSE; } else { if (!exec_ref_pic_marking_sliding_window(decoder)) return FALSE; } } return TRUE; } static void vaapi_init_picture(VAPictureH264 *pic) { pic->picture_id = VA_INVALID_ID; pic->frame_idx = 0; pic->flags = VA_PICTURE_H264_INVALID; pic->TopFieldOrderCnt = 0; pic->BottomFieldOrderCnt = 0; } static void vaapi_fill_picture(VAPictureH264 *pic, GstVaapiPictureH264 *picture, guint picture_structure) { if (!picture_structure) picture_structure = picture->structure; pic->picture_id = picture->base.surface_id; pic->flags = 0; if (GST_VAAPI_PICTURE_IS_LONG_TERM_REFERENCE(picture)) { pic->flags |= VA_PICTURE_H264_LONG_TERM_REFERENCE; pic->frame_idx = picture->long_term_frame_idx; } else { if (GST_VAAPI_PICTURE_IS_SHORT_TERM_REFERENCE(picture)) pic->flags |= VA_PICTURE_H264_SHORT_TERM_REFERENCE; pic->frame_idx = picture->frame_num; } switch (picture_structure) { case GST_VAAPI_PICTURE_STRUCTURE_FRAME: pic->TopFieldOrderCnt = picture->field_poc[TOP_FIELD]; pic->BottomFieldOrderCnt = picture->field_poc[BOTTOM_FIELD]; break; case GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD: pic->flags |= VA_PICTURE_H264_TOP_FIELD; pic->TopFieldOrderCnt = picture->field_poc[TOP_FIELD]; pic->BottomFieldOrderCnt = 0; break; case GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD: pic->flags |= VA_PICTURE_H264_BOTTOM_FIELD; pic->BottomFieldOrderCnt = picture->field_poc[BOTTOM_FIELD]; pic->TopFieldOrderCnt = 0; break; } } static gboolean fill_picture(GstVaapiDecoderH264 *decoder, GstVaapiPictureH264 *picture, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPicture * const base_picture = &picture->base; GstH264SliceHdr * const slice_hdr = &unit->data.slice_hdr; GstH264PPS * const pps = picture->pps; GstH264SPS * const sps = pps->sequence; VAPictureParameterBufferH264 * const pic_param = base_picture->param; guint i, n; /* Fill in VAPictureParameterBufferH264 */ vaapi_fill_picture(&pic_param->CurrPic, picture, 0); for (i = 0, n = 0; i < priv->dpb_count; i++) { GstVaapiFrameStore * const fs = priv->dpb[i]; if (gst_vaapi_frame_store_has_reference(fs)) vaapi_fill_picture(&pic_param->ReferenceFrames[n++], fs->buffers[0], fs->structure); } for (; n < G_N_ELEMENTS(pic_param->ReferenceFrames); n++) vaapi_init_picture(&pic_param->ReferenceFrames[n]); #define COPY_FIELD(s, f) \ pic_param->f = (s)->f #define COPY_BFM(a, s, f) \ pic_param->a.bits.f = (s)->f pic_param->picture_width_in_mbs_minus1 = priv->mb_width - 1; pic_param->picture_height_in_mbs_minus1 = priv->mb_height - 1; pic_param->frame_num = priv->frame_num; COPY_FIELD(sps, bit_depth_luma_minus8); COPY_FIELD(sps, bit_depth_chroma_minus8); COPY_FIELD(sps, num_ref_frames); COPY_FIELD(pps, num_slice_groups_minus1); COPY_FIELD(pps, slice_group_map_type); COPY_FIELD(pps, slice_group_change_rate_minus1); COPY_FIELD(pps, pic_init_qp_minus26); COPY_FIELD(pps, pic_init_qs_minus26); COPY_FIELD(pps, chroma_qp_index_offset); COPY_FIELD(pps, second_chroma_qp_index_offset); pic_param->seq_fields.value = 0; /* reset all bits */ pic_param->seq_fields.bits.residual_colour_transform_flag = sps->separate_colour_plane_flag; pic_param->seq_fields.bits.MinLumaBiPredSize8x8 = sps->level_idc >= 31; /* A.3.3.2 */ COPY_BFM(seq_fields, sps, chroma_format_idc); COPY_BFM(seq_fields, sps, gaps_in_frame_num_value_allowed_flag); COPY_BFM(seq_fields, sps, frame_mbs_only_flag); COPY_BFM(seq_fields, sps, mb_adaptive_frame_field_flag); COPY_BFM(seq_fields, sps, direct_8x8_inference_flag); COPY_BFM(seq_fields, sps, log2_max_frame_num_minus4); COPY_BFM(seq_fields, sps, pic_order_cnt_type); COPY_BFM(seq_fields, sps, log2_max_pic_order_cnt_lsb_minus4); COPY_BFM(seq_fields, sps, delta_pic_order_always_zero_flag); pic_param->pic_fields.value = 0; /* reset all bits */ pic_param->pic_fields.bits.field_pic_flag = slice_hdr->field_pic_flag; pic_param->pic_fields.bits.reference_pic_flag = GST_VAAPI_PICTURE_IS_REFERENCE(picture); COPY_BFM(pic_fields, pps, entropy_coding_mode_flag); COPY_BFM(pic_fields, pps, weighted_pred_flag); COPY_BFM(pic_fields, pps, weighted_bipred_idc); COPY_BFM(pic_fields, pps, transform_8x8_mode_flag); COPY_BFM(pic_fields, pps, constrained_intra_pred_flag); COPY_BFM(pic_fields, pps, pic_order_present_flag); COPY_BFM(pic_fields, pps, deblocking_filter_control_present_flag); COPY_BFM(pic_fields, pps, redundant_pic_cnt_present_flag); return TRUE; } /* Detection of the first VCL NAL unit of a primary coded picture (7.4.1.2.4) */ static gboolean is_new_picture(GstVaapiDecoderUnitH264 *unit, GstVaapiDecoderUnitH264 *prev_unit) { GstH264SliceHdr * const slice_hdr = &unit->data.slice_hdr; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; GstH264SliceHdr *prev_slice_hdr; if (!prev_unit) return TRUE; prev_slice_hdr = &prev_unit->data.slice_hdr; #define CHECK_EXPR(expr, field_name) do { \ if (!(expr)) { \ GST_DEBUG(field_name " differs in value"); \ return TRUE; \ } \ } while (0) #define CHECK_VALUE(new_slice_hdr, old_slice_hdr, field) \ CHECK_EXPR(((new_slice_hdr)->field == (old_slice_hdr)->field), #field) /* frame_num differs in value, regardless of inferred values to 0 */ CHECK_VALUE(slice_hdr, prev_slice_hdr, frame_num); /* pic_parameter_set_id differs in value */ CHECK_VALUE(slice_hdr, prev_slice_hdr, pps); /* field_pic_flag differs in value */ CHECK_VALUE(slice_hdr, prev_slice_hdr, field_pic_flag); /* bottom_field_flag is present in both and differs in value */ if (slice_hdr->field_pic_flag && prev_slice_hdr->field_pic_flag) CHECK_VALUE(slice_hdr, prev_slice_hdr, bottom_field_flag); /* nal_ref_idc differs in value with one of the nal_ref_idc values is 0 */ CHECK_EXPR((unit->nalu.ref_idc != 0) == (prev_unit->nalu.ref_idc != 0), "nal_ref_idc"); /* POC type is 0 for both and either pic_order_cnt_lsb differs in value or delta_pic_order_cnt_bottom differs in value */ if (sps->pic_order_cnt_type == 0) { CHECK_VALUE(slice_hdr, prev_slice_hdr, pic_order_cnt_lsb); if (pps->pic_order_present_flag && !slice_hdr->field_pic_flag) CHECK_VALUE(slice_hdr, prev_slice_hdr, delta_pic_order_cnt_bottom); } /* POC type is 1 for both and either delta_pic_order_cnt[0] differs in value or delta_pic_order_cnt[1] differs in value */ else if (sps->pic_order_cnt_type == 1) { CHECK_VALUE(slice_hdr, prev_slice_hdr, delta_pic_order_cnt[0]); CHECK_VALUE(slice_hdr, prev_slice_hdr, delta_pic_order_cnt[1]); } /* IdrPicFlag differs in value */ CHECK_VALUE(&unit->nalu, &prev_unit->nalu, idr_pic_flag); /* IdrPicFlag is equal to 1 for both and idr_pic_id differs in value */ if (unit->nalu.idr_pic_flag) CHECK_VALUE(slice_hdr, prev_slice_hdr, idr_pic_id); #undef CHECK_EXPR #undef CHECK_VALUE return FALSE; } static GstVaapiDecoderStatus decode_picture(GstVaapiDecoderH264 *decoder, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstH264SliceHdr * const slice_hdr = &unit->data.slice_hdr; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; GstVaapiPictureH264 *picture; GstVaapiDecoderStatus status; status = ensure_context(decoder, sps); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) return status; if (priv->current_picture) { /* Re-use current picture where the first field was decoded */ picture = gst_vaapi_picture_h264_new_field(priv->current_picture); if (!picture) { GST_ERROR("failed to allocate field picture"); return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED; } } else { /* Create new picture */ picture = gst_vaapi_picture_h264_new(decoder); if (!picture) { GST_ERROR("failed to allocate picture"); return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED; } } gst_vaapi_picture_replace(&priv->current_picture, picture); gst_vaapi_picture_unref(picture); picture->pps = pps; status = ensure_quant_matrix(decoder, picture); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) { GST_ERROR("failed to reset quantizer matrix"); return status; } if (!init_picture(decoder, picture, unit)) return GST_VAAPI_DECODER_STATUS_ERROR_UNKNOWN; if (!fill_picture(decoder, picture, unit)) return GST_VAAPI_DECODER_STATUS_ERROR_UNKNOWN; return GST_VAAPI_DECODER_STATUS_SUCCESS; } static inline guint get_slice_data_bit_offset(GstH264SliceHdr *slice_hdr) { guint epb_count; epb_count = slice_hdr->n_emulation_prevention_bytes; return 8 /* nal_unit_type */ + slice_hdr->header_size - epb_count * 8; } static gboolean fill_pred_weight_table(GstVaapiDecoderH264 *decoder, GstVaapiSlice *slice, GstH264SliceHdr *slice_hdr) { VASliceParameterBufferH264 * const slice_param = slice->param; GstH264PPS * const pps = slice_hdr->pps; GstH264SPS * const sps = pps->sequence; GstH264PredWeightTable * const w = &slice_hdr->pred_weight_table; guint num_weight_tables = 0; gint i, j; if (pps->weighted_pred_flag && (GST_H264_IS_P_SLICE(slice_hdr) || GST_H264_IS_SP_SLICE(slice_hdr))) num_weight_tables = 1; else if (pps->weighted_bipred_idc == 1 && GST_H264_IS_B_SLICE(slice_hdr)) num_weight_tables = 2; else num_weight_tables = 0; slice_param->luma_log2_weight_denom = w->luma_log2_weight_denom; slice_param->chroma_log2_weight_denom = w->chroma_log2_weight_denom; slice_param->luma_weight_l0_flag = 0; slice_param->chroma_weight_l0_flag = 0; slice_param->luma_weight_l1_flag = 0; slice_param->chroma_weight_l1_flag = 0; if (num_weight_tables < 1) return TRUE; slice_param->luma_weight_l0_flag = 1; for (i = 0; i <= slice_param->num_ref_idx_l0_active_minus1; i++) { slice_param->luma_weight_l0[i] = w->luma_weight_l0[i]; slice_param->luma_offset_l0[i] = w->luma_offset_l0[i]; } slice_param->chroma_weight_l0_flag = sps->chroma_array_type != 0; if (slice_param->chroma_weight_l0_flag) { for (i = 0; i <= slice_param->num_ref_idx_l0_active_minus1; i++) { for (j = 0; j < 2; j++) { slice_param->chroma_weight_l0[i][j] = w->chroma_weight_l0[i][j]; slice_param->chroma_offset_l0[i][j] = w->chroma_offset_l0[i][j]; } } } if (num_weight_tables < 2) return TRUE; slice_param->luma_weight_l1_flag = 1; for (i = 0; i <= slice_param->num_ref_idx_l1_active_minus1; i++) { slice_param->luma_weight_l1[i] = w->luma_weight_l1[i]; slice_param->luma_offset_l1[i] = w->luma_offset_l1[i]; } slice_param->chroma_weight_l1_flag = sps->chroma_array_type != 0; if (slice_param->chroma_weight_l1_flag) { for (i = 0; i <= slice_param->num_ref_idx_l1_active_minus1; i++) { for (j = 0; j < 2; j++) { slice_param->chroma_weight_l1[i][j] = w->chroma_weight_l1[i][j]; slice_param->chroma_offset_l1[i][j] = w->chroma_offset_l1[i][j]; } } } return TRUE; } static gboolean fill_RefPicList(GstVaapiDecoderH264 *decoder, GstVaapiSlice *slice, GstH264SliceHdr *slice_hdr) { GstVaapiDecoderH264Private * const priv = decoder->priv; VASliceParameterBufferH264 * const slice_param = slice->param; guint i, num_ref_lists = 0; slice_param->num_ref_idx_l0_active_minus1 = 0; slice_param->num_ref_idx_l1_active_minus1 = 0; if (GST_H264_IS_B_SLICE(slice_hdr)) num_ref_lists = 2; else if (GST_H264_IS_I_SLICE(slice_hdr)) num_ref_lists = 0; else num_ref_lists = 1; if (num_ref_lists < 1) return TRUE; slice_param->num_ref_idx_l0_active_minus1 = slice_hdr->num_ref_idx_l0_active_minus1; for (i = 0; i < priv->RefPicList0_count && priv->RefPicList0[i]; i++) vaapi_fill_picture(&slice_param->RefPicList0[i], priv->RefPicList0[i], 0); for (; i <= slice_param->num_ref_idx_l0_active_minus1; i++) vaapi_init_picture(&slice_param->RefPicList0[i]); if (num_ref_lists < 2) return TRUE; slice_param->num_ref_idx_l1_active_minus1 = slice_hdr->num_ref_idx_l1_active_minus1; for (i = 0; i < priv->RefPicList1_count && priv->RefPicList1[i]; i++) vaapi_fill_picture(&slice_param->RefPicList1[i], priv->RefPicList1[i], 0); for (; i <= slice_param->num_ref_idx_l1_active_minus1; i++) vaapi_init_picture(&slice_param->RefPicList1[i]); return TRUE; } static gboolean fill_slice(GstVaapiDecoderH264 *decoder, GstVaapiSlice *slice, GstH264SliceHdr *slice_hdr) { VASliceParameterBufferH264 * const slice_param = slice->param; /* Fill in VASliceParameterBufferH264 */ slice_param->slice_data_bit_offset = get_slice_data_bit_offset(slice_hdr); slice_param->first_mb_in_slice = slice_hdr->first_mb_in_slice; slice_param->slice_type = slice_hdr->type % 5; slice_param->direct_spatial_mv_pred_flag = slice_hdr->direct_spatial_mv_pred_flag; slice_param->cabac_init_idc = slice_hdr->cabac_init_idc; slice_param->slice_qp_delta = slice_hdr->slice_qp_delta; slice_param->disable_deblocking_filter_idc = slice_hdr->disable_deblocking_filter_idc; slice_param->slice_alpha_c0_offset_div2 = slice_hdr->slice_alpha_c0_offset_div2; slice_param->slice_beta_offset_div2 = slice_hdr->slice_beta_offset_div2; if (!fill_RefPicList(decoder, slice, slice_hdr)) return FALSE; if (!fill_pred_weight_table(decoder, slice, slice_hdr)) return FALSE; return TRUE; } static GstVaapiDecoderStatus decode_slice(GstVaapiDecoderH264 *decoder, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiPictureH264 * const picture = priv->current_picture; GstH264SliceHdr * const slice_hdr = &unit->data.slice_hdr; GstH264NalUnit * const nalu = &unit->nalu; GstVaapiSlice *slice; GST_DEBUG("slice (%u bytes)", nalu->size); if (!priv->got_sps || !priv->got_pps) { GST_ERROR("not initialized yet"); return GST_VAAPI_DECODER_STATUS_SUCCESS; } unit->base.buffer = gst_buffer_create_sub( GST_VAAPI_DECODER_CODEC_FRAME(decoder)->input_buffer, unit->base.offset, unit->base.size); if (!unit->base.buffer) { GST_ERROR("failed to allocate slice data"); return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED; } slice = GST_VAAPI_SLICE_NEW(H264, decoder, GST_BUFFER_DATA(unit->base.buffer) + nalu->offset, nalu->size); if (!slice) { GST_ERROR("failed to allocate slice"); return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED; } if (!fill_slice(decoder, slice, slice_hdr)) { gst_vaapi_mini_object_unref(GST_VAAPI_MINI_OBJECT(slice)); return GST_VAAPI_DECODER_STATUS_ERROR_UNKNOWN; } gst_vaapi_picture_add_slice(GST_VAAPI_PICTURE_CAST(picture), slice); picture->last_slice_hdr = slice_hdr; return GST_VAAPI_DECODER_STATUS_SUCCESS; } static inline gint scan_for_start_code(GstAdapter *adapter, guint ofs, guint size, guint32 *scp) { return (gint)gst_adapter_masked_scan_uint32_peek(adapter, 0xffffff00, 0x00000100, ofs, size, scp); } static GstVaapiDecoderStatus decode_unit(GstVaapiDecoderH264 *decoder, GstVaapiDecoderUnitH264 *unit) { GstVaapiDecoderStatus status; switch (unit->nalu.type) { case GST_H264_NAL_SLICE_IDR: /* fall-through. IDR specifics are handled in init_picture() */ case GST_H264_NAL_SLICE: status = decode_slice(decoder, unit); break; case GST_H264_NAL_SEQ_END: status = decode_sequence_end(decoder); break; case GST_H264_NAL_SEI: status = GST_VAAPI_DECODER_STATUS_SUCCESS; break; default: GST_WARNING("unsupported NAL unit type %d", unit->nalu.type); status = GST_VAAPI_DECODER_STATUS_ERROR_BITSTREAM_PARSER; break; } return status; } static GstVaapiDecoderStatus decode_codec_data(GstVaapiDecoderH264 *decoder, GstBuffer *buffer) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiDecoderStatus status; GstVaapiDecoderUnitH264 unit; GstH264ParserResult result; guchar *buf; guint buf_size; guint i, ofs, num_sps, num_pps; buf = GST_BUFFER_DATA(buffer); buf_size = GST_BUFFER_SIZE(buffer); if (!buf || buf_size == 0) return GST_VAAPI_DECODER_STATUS_SUCCESS; if (buf_size < 8) return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA; if (buf[0] != 1) { GST_ERROR("failed to decode codec-data, not in avcC format"); return GST_VAAPI_DECODER_STATUS_ERROR_BITSTREAM_PARSER; } priv->nal_length_size = (buf[4] & 0x03) + 1; num_sps = buf[5] & 0x1f; ofs = 6; for (i = 0; i < num_sps; i++) { result = gst_h264_parser_identify_nalu_avc( priv->parser, buf, ofs, buf_size, 2, &unit.nalu ); if (result != GST_H264_PARSER_OK) return get_status(result); status = parse_sps(decoder, &unit); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) return status; ofs = unit.nalu.offset + unit.nalu.size; } num_pps = buf[ofs]; ofs++; for (i = 0; i < num_pps; i++) { result = gst_h264_parser_identify_nalu_avc( priv->parser, buf, ofs, buf_size, 2, &unit.nalu ); if (result != GST_H264_PARSER_OK) return get_status(result); status = parse_pps(decoder, &unit); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) return status; ofs = unit.nalu.offset + unit.nalu.size; } priv->is_avcC = TRUE; return GST_VAAPI_DECODER_STATUS_SUCCESS; } static GstVaapiDecoderStatus ensure_decoder(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiDecoderStatus status; GstBuffer *codec_data; g_return_val_if_fail(priv->is_constructed, GST_VAAPI_DECODER_STATUS_ERROR_INIT_FAILED); if (!priv->is_opened) { priv->is_opened = gst_vaapi_decoder_h264_open(decoder); if (!priv->is_opened) return GST_VAAPI_DECODER_STATUS_ERROR_UNSUPPORTED_CODEC; codec_data = GST_VAAPI_DECODER_CODEC_DATA(decoder); if (codec_data) { status = decode_codec_data(decoder, codec_data); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) return status; } } return GST_VAAPI_DECODER_STATUS_SUCCESS; } static GstVaapiDecoderStatus gst_vaapi_decoder_h264_parse(GstVaapiDecoder *base_decoder, GstAdapter *adapter, gboolean at_eos, GstVaapiDecoderUnit **unit_ptr) { GstVaapiDecoderH264 * const decoder = GST_VAAPI_DECODER_H264_CAST(base_decoder); GstVaapiDecoderH264Private * const priv = decoder->priv; GstVaapiParserState * const ps = GST_VAAPI_PARSER_STATE(base_decoder); GstVaapiDecoderUnitH264 *unit; GstVaapiDecoderStatus status; GstH264ParserResult result; guchar *buf; guint i, size, buf_size, nalu_size, flags; guint32 start_code; gint ofs, ofs2; status = ensure_decoder(decoder); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) return status; size = gst_adapter_available(adapter); if (priv->is_avcC) { if (size < priv->nal_length_size) return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA; buf = (guchar *)&start_code; g_assert(priv->nal_length_size <= sizeof(start_code)); gst_adapter_copy(adapter, buf, 0, priv->nal_length_size); nalu_size = 0; for (i = 0; i < priv->nal_length_size; i++) nalu_size = (nalu_size << 8) | buf[i]; buf_size = priv->nal_length_size + nalu_size; if (size < buf_size) return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA; } else { if (size < 4) return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA; ofs = scan_for_start_code(adapter, 0, size, NULL); if (ofs < 0) return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA; if (ofs > 0) { gst_adapter_flush(adapter, ofs); size -= ofs; } ofs2 = ps->input_offset2 - ofs - 4; if (ofs2 < 4) ofs2 = 4; ofs = G_UNLIKELY(size < ofs2 + 4) ? -1 : scan_for_start_code(adapter, ofs2, size - ofs2, NULL); if (ofs < 0) { // Assume the whole NAL unit is present if end-of-stream if (!at_eos) { ps->input_offset2 = size; return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA; } ofs = size; } buf_size = ofs; } ps->input_offset2 = 0; buf = (guchar *)gst_adapter_peek(adapter, buf_size); if (!buf) return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA; unit = gst_vaapi_decoder_unit_h264_new(buf_size); if (!unit) return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED; if (priv->is_avcC) result = gst_h264_parser_identify_nalu_avc(priv->parser, buf, 0, buf_size, priv->nal_length_size, &unit->nalu); else result = gst_h264_parser_identify_nalu_unchecked(priv->parser, buf, 0, buf_size, &unit->nalu); status = get_status(result); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) { gst_vaapi_decoder_unit_unref(unit); return status; } switch (unit->nalu.type) { case GST_H264_NAL_SPS: status = parse_sps(decoder, unit); break; case GST_H264_NAL_PPS: status = parse_pps(decoder, unit); break; case GST_H264_NAL_SEI: status = parse_sei(decoder, unit); break; case GST_H264_NAL_SLICE_IDR: case GST_H264_NAL_SLICE: status = parse_slice(decoder, unit); break; default: status = GST_VAAPI_DECODER_STATUS_SUCCESS; break; } if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) return status; flags = 0; switch (unit->nalu.type) { case GST_H264_NAL_AU_DELIMITER: flags |= GST_VAAPI_DECODER_UNIT_FLAG_FRAME_START; /* fall-through */ case GST_H264_NAL_FILLER_DATA: flags |= GST_VAAPI_DECODER_UNIT_FLAG_SKIP; break; case GST_H264_NAL_STREAM_END: flags |= GST_VAAPI_DECODER_UNIT_FLAG_STREAM_END; /* fall-through */ case GST_H264_NAL_SEQ_END: flags |= GST_VAAPI_DECODER_UNIT_FLAG_FRAME_END; break; case GST_H264_NAL_SPS: case GST_H264_NAL_PPS: flags |= GST_VAAPI_DECODER_UNIT_FLAG_SKIP; /* fall-through */ case GST_H264_NAL_SEI: flags |= GST_VAAPI_DECODER_UNIT_FLAG_FRAME_START; break; case GST_H264_NAL_SLICE_IDR: case GST_H264_NAL_SLICE: flags |= GST_VAAPI_DECODER_UNIT_FLAG_SLICE; if (is_new_picture(unit, priv->prev_slice_unit)) flags |= GST_VAAPI_DECODER_UNIT_FLAG_FRAME_START; gst_vaapi_decoder_unit_replace(&priv->prev_slice_unit, unit); break; default: if (unit->nalu.type >= 14 && unit->nalu.type <= 18) flags |= GST_VAAPI_DECODER_UNIT_FLAG_FRAME_START; break; } GST_VAAPI_DECODER_UNIT_FLAG_SET(unit, flags); unit->nalu.data = NULL; *unit_ptr = &unit->base; return GST_VAAPI_DECODER_STATUS_SUCCESS; } static GstVaapiDecoderStatus gst_vaapi_decoder_h264_decode(GstVaapiDecoder *base_decoder, GstVaapiDecoderUnit *unit) { GstVaapiDecoderH264 * const decoder = GST_VAAPI_DECODER_H264_CAST(base_decoder); GstVaapiDecoderStatus status; status = ensure_decoder(decoder); if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) return status; return decode_unit(decoder, (GstVaapiDecoderUnitH264 *)unit); } static GstVaapiDecoderStatus gst_vaapi_decoder_h264_start_frame(GstVaapiDecoder *base_decoder, GstVaapiDecoderUnit *base_unit) { GstVaapiDecoderH264 * const decoder = GST_VAAPI_DECODER_H264_CAST(base_decoder); GstVaapiDecoderUnitH264 * const unit = (GstVaapiDecoderUnitH264 *)base_unit; return decode_picture(decoder, unit); } static GstVaapiDecoderStatus gst_vaapi_decoder_h264_end_frame(GstVaapiDecoder *base_decoder) { GstVaapiDecoderH264 * const decoder = GST_VAAPI_DECODER_H264_CAST(base_decoder); return decode_current_picture(decoder); } static void gst_vaapi_decoder_h264_finalize(GObject *object) { GstVaapiDecoderH264 * const decoder = GST_VAAPI_DECODER_H264_CAST(object); gst_vaapi_decoder_h264_destroy(decoder); G_OBJECT_CLASS(gst_vaapi_decoder_h264_parent_class)->finalize(object); } static void gst_vaapi_decoder_h264_constructed(GObject *object) { GstVaapiDecoderH264 * const decoder = GST_VAAPI_DECODER_H264_CAST(object); GstVaapiDecoderH264Private * const priv = decoder->priv; GObjectClass *parent_class; parent_class = G_OBJECT_CLASS(gst_vaapi_decoder_h264_parent_class); if (parent_class->constructed) parent_class->constructed(object); priv->is_constructed = gst_vaapi_decoder_h264_create(decoder); } static void gst_vaapi_decoder_h264_class_init(GstVaapiDecoderH264Class *klass) { GObjectClass * const object_class = G_OBJECT_CLASS(klass); GstVaapiDecoderClass * const decoder_class = GST_VAAPI_DECODER_CLASS(klass); g_type_class_add_private(klass, sizeof(GstVaapiDecoderH264Private)); object_class->finalize = gst_vaapi_decoder_h264_finalize; object_class->constructed = gst_vaapi_decoder_h264_constructed; decoder_class->parse = gst_vaapi_decoder_h264_parse; decoder_class->decode = gst_vaapi_decoder_h264_decode; decoder_class->start_frame = gst_vaapi_decoder_h264_start_frame; decoder_class->end_frame = gst_vaapi_decoder_h264_end_frame; } static void gst_vaapi_decoder_h264_init(GstVaapiDecoderH264 *decoder) { GstVaapiDecoderH264Private *priv; priv = GST_VAAPI_DECODER_H264_GET_PRIVATE(decoder); decoder->priv = priv; priv->profile = GST_VAAPI_PROFILE_UNKNOWN; priv->entrypoint = GST_VAAPI_ENTRYPOINT_VLD; priv->chroma_type = GST_VAAPI_CHROMA_TYPE_YUV420; priv->prev_pic_structure = GST_VAAPI_PICTURE_STRUCTURE_FRAME; priv->progressive_sequence = TRUE; } /** * gst_vaapi_decoder_h264_new: * @display: a #GstVaapiDisplay * @caps: a #GstCaps holding codec information * * Creates a new #GstVaapiDecoder for MPEG-2 decoding. The @caps can * hold extra information like codec-data and pictured coded size. * * Return value: the newly allocated #GstVaapiDecoder object */ GstVaapiDecoder * gst_vaapi_decoder_h264_new(GstVaapiDisplay *display, GstCaps *caps) { GstVaapiDecoderH264 *decoder; g_return_val_if_fail(GST_VAAPI_IS_DISPLAY(display), NULL); g_return_val_if_fail(GST_IS_CAPS(caps), NULL); decoder = g_object_new( GST_VAAPI_TYPE_DECODER_H264, "display", display, "caps", caps, NULL ); if (!decoder->priv->is_constructed) { g_object_unref(decoder); return NULL; } return GST_VAAPI_DECODER_CAST(decoder); }