/* GStreamer * Copyright (C) 2020 Igalia, S.L. * Author: Víctor Jáquez * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the0 * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /** * SECTION:element-vavp9dec * @title: vavp9dec * @short_description: A VA-API based VP9 video decoder * * vavp9dec decodes VP9 bitstreams to VA surfaces using the * installed and chosen [VA-API](https://01.org/linuxmedia/vaapi) * driver. * * The decoding surfaces can be mapped onto main memory as video * frames. * * ## Example launch line * ``` * gst-launch-1.0 filesrc location=sample.webm ! parsebin ! vavp9dec ! autovideosink * ``` * * Since: 1.20 * */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "gstvavp9dec.h" #include "gstvabasedec.h" GST_DEBUG_CATEGORY_STATIC (gst_va_vp9dec_debug); #ifndef GST_DISABLE_GST_DEBUG #define GST_CAT_DEFAULT gst_va_vp9dec_debug #else #define GST_CAT_DEFAULT NULL #endif #define GST_VA_VP9_DEC(obj) ((GstVaVp9Dec *) obj) #define GST_VA_VP9_DEC_GET_CLASS(obj) (G_TYPE_INSTANCE_GET_CLASS ((obj), G_TYPE_FROM_INSTANCE (obj), GstVaVp9DecClass)) #define GST_VA_VP9_DEC_CLASS(klass) ((GstVaVp9DecClass *) klass) typedef struct _GstVaVp9Dec GstVaVp9Dec; typedef struct _GstVaVp9DecClass GstVaVp9DecClass; struct _GstVaVp9DecClass { GstVaBaseDecClass parent_class; }; struct _GstVaVp9Dec { GstVaBaseDec parent; GstVp9Segmentation segmentation[GST_VP9_MAX_SEGMENTS]; gboolean need_negotiation; }; /* *INDENT-OFF* */ static const gchar *src_caps_str = GST_VIDEO_CAPS_MAKE_WITH_FEATURES (GST_CAPS_FEATURE_MEMORY_VA, "{ NV12 }") " ;" GST_VIDEO_CAPS_MAKE ("{ NV12 }"); /* *INDENT-ON* */ static const gchar *sink_caps_str = "video/x-vp9"; static guint _get_rtformat (GstVaVp9Dec * self, GstVP9Profile profile, GstVp9BitDepth bit_depth, gint subsampling_x, gint subsampling_y) { switch (profile) { case GST_VP9_PROFILE_0: return VA_RT_FORMAT_YUV420; case GST_VP9_PROFILE_1: if (subsampling_x == 1 && subsampling_y == 0) return VA_RT_FORMAT_YUV422; else if (subsampling_x == 0 && subsampling_y == 0) return VA_RT_FORMAT_YUV444; break; case GST_VP9_PROFILE_2: if (bit_depth == GST_VP9_BIT_DEPTH_10) return VA_RT_FORMAT_YUV420_10; else if (bit_depth == GST_VP9_BIT_DEPTH_12) return VA_RT_FORMAT_YUV420_12; break; case GST_VP9_PROFILE_3: if (subsampling_x == 1 && subsampling_y == 0) { if (bit_depth == GST_VP9_BIT_DEPTH_10) return VA_RT_FORMAT_YUV422_10; else if (bit_depth == GST_VP9_BIT_DEPTH_12) return VA_RT_FORMAT_YUV422_12; } else if (subsampling_x == 0 && subsampling_y == 0) { if (bit_depth == GST_VP9_BIT_DEPTH_10) return VA_RT_FORMAT_YUV444_10; else if (bit_depth == GST_VP9_BIT_DEPTH_12) return VA_RT_FORMAT_YUV444_12; } break; default: break; } GST_ERROR_OBJECT (self, "Unsupported chroma format"); return 0; } static VAProfile _get_profile (GstVaVp9Dec * self, GstVP9Profile profile) { switch (profile) { case GST_VP9_PROFILE_0: return VAProfileVP9Profile0; case GST_VP9_PROFILE_1: return VAProfileVP9Profile1; case GST_VP9_PROFILE_2: return VAProfileVP9Profile2; case GST_VP9_PROFILE_3: return VAProfileVP9Profile3; default: break; } GST_ERROR_OBJECT (self, "Unsupported profile"); return VAProfileNone; } static gboolean gst_va_vp9_new_sequence (GstVp9Decoder * decoder, const GstVp9FrameHeader * frame_hdr) { GstVaBaseDec *base = GST_VA_BASE_DEC (decoder); GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder); VAProfile profile; gboolean negotiation_needed = FALSE; guint rt_format; profile = _get_profile (self, frame_hdr->profile); if (profile == VAProfileNone) return FALSE; if (!gst_va_decoder_has_profile (base->decoder, profile)) { GST_ERROR_OBJECT (self, "Profile %s is not supported", gst_va_profile_name (profile)); return FALSE; } rt_format = _get_rtformat (self, frame_hdr->profile, frame_hdr->bit_depth, frame_hdr->subsampling_x, frame_hdr->subsampling_y); if (rt_format == 0) return FALSE; if (!gst_va_decoder_config_is_equal (base->decoder, profile, rt_format, frame_hdr->width, frame_hdr->height)) { base->profile = profile; base->width = frame_hdr->width; base->height = frame_hdr->height; base->rt_format = rt_format; negotiation_needed = TRUE; } base->min_buffers = GST_VP9_REF_FRAMES; if (negotiation_needed) { self->need_negotiation = TRUE; if (!gst_video_decoder_negotiate (GST_VIDEO_DECODER (self))) { GST_ERROR_OBJECT (self, "Failed to negotiate with downstream"); return FALSE; } } return TRUE; } static gboolean _check_resolution_change (GstVaVp9Dec * self, GstVp9Picture * picture) { GstVaBaseDec *base = GST_VA_BASE_DEC (self); const GstVp9FrameHeader *frame_hdr = &picture->frame_hdr; if ((base->width != frame_hdr->width) || base->height != frame_hdr->height) { base->width = frame_hdr->width; base->height = frame_hdr->height; self->need_negotiation = TRUE; if (!gst_video_decoder_negotiate (GST_VIDEO_DECODER (self))) { GST_ERROR_OBJECT (self, "Resolution changed, but failed to" " negotiate with downstream"); return FALSE; /* @TODO: if negotiation fails, decoder should resize output * frame. For that we would need an auxiliar allocator, and * later use GstVaFilter or GstVideoConverter. */ } } return TRUE; } static gboolean gst_va_vp9_dec_new_picture (GstVp9Decoder * decoder, GstVideoCodecFrame * frame, GstVp9Picture * picture) { GstFlowReturn ret; GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder); GstVaDecodePicture *pic; GstVideoDecoder *vdec = GST_VIDEO_DECODER (decoder); GstVaBaseDec *base = GST_VA_BASE_DEC (decoder); if (!_check_resolution_change (self, picture)) return FALSE; ret = gst_video_decoder_allocate_output_frame (vdec, frame); if (ret != GST_FLOW_OK) goto error; pic = gst_va_decode_picture_new (base->decoder, frame->output_buffer); gst_vp9_picture_set_user_data (picture, pic, (GDestroyNotify) gst_va_decode_picture_free); GST_LOG_OBJECT (self, "New va decode picture %p - %#x", pic, gst_va_decode_picture_get_surface (pic)); return TRUE; error: { GST_WARNING_OBJECT (self, "Failed to allocated output buffer, return %s", gst_flow_get_name (ret)); return FALSE; } } static inline gboolean _fill_param (GstVp9Decoder * decoder, GstVp9Picture * picture, GstVp9Dpb * dpb) { GstVaBaseDec *base = GST_VA_BASE_DEC (decoder); GstVaDecodePicture *va_pic; const GstVp9FrameHeader *frame_hdr = &picture->frame_hdr; const GstVp9LoopFilterParams *lfp = &frame_hdr->loop_filter_params; const GstVp9SegmentationParams *sp = &frame_hdr->segmentation_params; VADecPictureParameterBufferVP9 pic_param; guint i; /* *INDENT-OFF* */ pic_param = (VADecPictureParameterBufferVP9) { .frame_width = base->width, .frame_height = base->height, .pic_fields.bits = { .subsampling_x = frame_hdr->subsampling_x, .subsampling_y = frame_hdr->subsampling_x, .frame_type = frame_hdr->frame_type, .show_frame = frame_hdr->show_frame, .error_resilient_mode = frame_hdr->error_resilient_mode, .intra_only = frame_hdr->intra_only, .allow_high_precision_mv = frame_hdr->allow_high_precision_mv, .mcomp_filter_type = frame_hdr->interpolation_filter, .frame_parallel_decoding_mode = frame_hdr->frame_parallel_decoding_mode, .reset_frame_context = frame_hdr->reset_frame_context, .refresh_frame_context = frame_hdr->refresh_frame_context, .frame_context_idx = frame_hdr->frame_context_idx, .segmentation_enabled = sp->segmentation_enabled, .segmentation_temporal_update = sp->segmentation_temporal_update, .segmentation_update_map = sp->segmentation_update_map, .last_ref_frame = frame_hdr->ref_frame_idx[GST_VP9_REF_FRAME_LAST - 1], .last_ref_frame_sign_bias = frame_hdr->ref_frame_sign_bias[GST_VP9_REF_FRAME_LAST], .golden_ref_frame = frame_hdr->ref_frame_idx[GST_VP9_REF_FRAME_GOLDEN - 1], .golden_ref_frame_sign_bias = frame_hdr->ref_frame_sign_bias[GST_VP9_REF_FRAME_GOLDEN], .alt_ref_frame = frame_hdr->ref_frame_idx[GST_VP9_REF_FRAME_ALTREF - 1], .alt_ref_frame_sign_bias = frame_hdr->ref_frame_sign_bias[GST_VP9_REF_FRAME_ALTREF], .lossless_flag = frame_hdr->lossless_flag, }, .filter_level = lfp->loop_filter_level, .sharpness_level = lfp->loop_filter_sharpness, .log2_tile_rows = frame_hdr->tile_rows_log2, .log2_tile_columns = frame_hdr->tile_cols_log2, .frame_header_length_in_bytes = frame_hdr->frame_header_length_in_bytes, .first_partition_size = frame_hdr->header_size_in_bytes, .profile = frame_hdr->profile, .bit_depth = frame_hdr->bit_depth }; /* *INDENT-ON* */ memcpy (pic_param.mb_segment_tree_probs, sp->segmentation_tree_probs, sizeof (sp->segmentation_tree_probs)); if (sp->segmentation_temporal_update) { memcpy (pic_param.segment_pred_probs, sp->segmentation_pred_prob, sizeof (sp->segmentation_pred_prob)); } else { memset (pic_param.segment_pred_probs, 255, sizeof (pic_param.segment_pred_probs)); } for (i = 0; i < GST_VP9_REF_FRAMES; i++) { if (dpb->pic_list[i]) { GstVaDecodePicture *va_pic = gst_vp9_picture_get_user_data (dpb->pic_list[i]); pic_param.reference_frames[i] = gst_va_decode_picture_get_surface (va_pic); } else { pic_param.reference_frames[i] = VA_INVALID_ID; } } va_pic = gst_vp9_picture_get_user_data (picture); return gst_va_decoder_add_param_buffer (base->decoder, va_pic, VAPictureParameterBufferType, &pic_param, sizeof (pic_param)); } static void _update_segmentation (GstVaVp9Dec * self, GstVp9FrameHeader * header) { const GstVp9LoopFilterParams *lfp = &header->loop_filter_params; const GstVp9QuantizationParams *qp = &header->quantization_params; const GstVp9SegmentationParams *sp = &header->segmentation_params; guint8 n_shift = lfp->loop_filter_level >> 5; guint i; for (i = 0; i < GST_VP9_MAX_SEGMENTS; i++) { gint16 luma_dc_quant_scale; gint16 luma_ac_quant_scale; gint16 chroma_dc_quant_scale; gint16 chroma_ac_quant_scale; guint8 qindex; guint8 lvl_lookup[GST_VP9_MAX_REF_LF_DELTAS][GST_VP9_MAX_MODE_LF_DELTAS]; gint lvl_seg = lfp->loop_filter_level; /* 8.6.1 Dequantization functions */ qindex = gst_vp9_get_qindex (sp, qp, i); luma_dc_quant_scale = gst_vp9_get_dc_quant (qindex, qp->delta_q_y_dc, header->bit_depth); luma_ac_quant_scale = gst_vp9_get_ac_quant (qindex, 0, header->bit_depth); chroma_dc_quant_scale = gst_vp9_get_dc_quant (qindex, qp->delta_q_uv_dc, header->bit_depth); chroma_ac_quant_scale = gst_vp9_get_ac_quant (qindex, qp->delta_q_uv_ac, header->bit_depth); if (!lfp->loop_filter_level) { memset (lvl_lookup, 0, sizeof (lvl_lookup)); } else { /* 8.8.1 Loop filter frame init process */ if (gst_vp9_seg_feature_active (sp, i, GST_VP9_SEG_LVL_ALT_L)) { if (sp->segmentation_abs_or_delta_update) { lvl_seg = sp->feature_data[i][GST_VP9_SEG_LVL_ALT_L]; } else { lvl_seg += sp->feature_data[i][GST_VP9_SEG_LVL_ALT_L]; } lvl_seg = CLAMP (lvl_seg, 0, GST_VP9_MAX_LOOP_FILTER); } if (!lfp->loop_filter_delta_enabled) { memset (lvl_lookup, lvl_seg, sizeof (lvl_lookup)); } else { guint8 ref, mode; gint intra_lvl = lvl_seg + (lfp->loop_filter_ref_deltas[GST_VP9_REF_FRAME_INTRA] << n_shift); memcpy (lvl_lookup, self->segmentation[i].filter_level, sizeof (lvl_lookup)); lvl_lookup[GST_VP9_REF_FRAME_INTRA][0] = CLAMP (intra_lvl, 0, GST_VP9_MAX_LOOP_FILTER); for (ref = GST_VP9_REF_FRAME_LAST; ref < GST_VP9_REF_FRAME_MAX; ref++) { for (mode = 0; mode < GST_VP9_MAX_MODE_LF_DELTAS; mode++) { intra_lvl = lvl_seg + (lfp->loop_filter_ref_deltas[ref] << n_shift) + (lfp->loop_filter_mode_deltas[mode] << n_shift); lvl_lookup[ref][mode] = CLAMP (intra_lvl, 0, GST_VP9_MAX_LOOP_FILTER); } } } } /* *INDENT-OFF* */ self->segmentation[i] = (GstVp9Segmentation) { .luma_dc_quant_scale = luma_dc_quant_scale, .luma_ac_quant_scale = luma_ac_quant_scale, .chroma_dc_quant_scale = chroma_dc_quant_scale, .chroma_ac_quant_scale = chroma_ac_quant_scale, .reference_frame_enabled = sp->feature_enabled[i][GST_VP9_SEG_LVL_REF_FRAME], .reference_frame = sp->feature_data[i][GST_VP9_SEG_LVL_REF_FRAME], .reference_skip = sp->feature_enabled[i][GST_VP9_SEG_SEG_LVL_SKIP], }; /* *INDENT-ON* */ memcpy (self->segmentation[i].filter_level, lvl_lookup, sizeof (lvl_lookup)); } } static inline gboolean _fill_slice (GstVp9Decoder * decoder, GstVp9Picture * picture) { GstVaBaseDec *base = GST_VA_BASE_DEC (decoder); GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder); GstVaDecodePicture *va_pic; const GstVp9Segmentation *seg; VASliceParameterBufferVP9 slice_param; guint i; _update_segmentation (self, &picture->frame_hdr); /* *INDENT-OFF* */ slice_param = (VASliceParameterBufferVP9) { .slice_data_size = picture->size, .slice_data_offset = 0, .slice_data_flag = VA_SLICE_DATA_FLAG_ALL, }; /* *INDENT-ON* */ for (i = 0; i < GST_VP9_MAX_SEGMENTS; i++) { seg = &self->segmentation[i]; /* *INDENT-OFF* */ slice_param.seg_param[i] = (VASegmentParameterVP9) { .segment_flags.fields = { .segment_reference_enabled = seg->reference_frame_enabled, .segment_reference = seg->reference_frame, .segment_reference_skipped = seg->reference_skip, }, .luma_dc_quant_scale = seg->luma_dc_quant_scale, .luma_ac_quant_scale = seg->luma_ac_quant_scale, .chroma_dc_quant_scale = seg->chroma_dc_quant_scale, .chroma_ac_quant_scale = seg->chroma_ac_quant_scale, }; /* *INDENT-ON* */ memcpy (slice_param.seg_param[i].filter_level, seg->filter_level, sizeof (slice_param.seg_param[i].filter_level)); } va_pic = gst_vp9_picture_get_user_data (picture); return gst_va_decoder_add_slice_buffer (base->decoder, va_pic, &slice_param, sizeof (slice_param), (gpointer) picture->data, picture->size); } static gboolean gst_va_vp9_decode_picture (GstVp9Decoder * decoder, GstVp9Picture * picture, GstVp9Dpb * dpb) { return _fill_param (decoder, picture, dpb) && _fill_slice (decoder, picture); } static gboolean gst_va_vp9_dec_end_picture (GstVp9Decoder * decoder, GstVp9Picture * picture) { GstVaBaseDec *base = GST_VA_BASE_DEC (decoder); GstVaDecodePicture *va_pic; GST_LOG_OBJECT (base, "end picture %p", picture); va_pic = gst_vp9_picture_get_user_data (picture); return gst_va_decoder_decode (base->decoder, va_pic); } static GstFlowReturn gst_va_vp9_dec_output_picture (GstVp9Decoder * decoder, GstVideoCodecFrame * frame, GstVp9Picture * picture) { GstVaBaseDec *base = GST_VA_BASE_DEC (decoder); GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder); GST_LOG_OBJECT (self, "Outputting picture %p", picture); if (base->copy_frames) gst_va_base_dec_copy_output_buffer (base, frame); gst_vp9_picture_unref (picture); return gst_video_decoder_finish_frame (GST_VIDEO_DECODER (self), frame); } static GstVp9Picture * gst_va_vp9_dec_duplicate_picture (GstVp9Decoder * decoder, GstVideoCodecFrame * frame, GstVp9Picture * picture) { GstVaDecodePicture *va_pic, *va_dup; GstVp9Picture *new_picture; if (!_check_resolution_change (GST_VA_VP9_DEC (decoder), picture)) return NULL; va_pic = gst_vp9_picture_get_user_data (picture); va_dup = gst_va_decode_picture_dup (va_pic); new_picture = gst_vp9_picture_new (); new_picture->frame_hdr = picture->frame_hdr; frame->output_buffer = gst_buffer_ref (va_dup->gstbuffer); gst_vp9_picture_set_user_data (picture, va_dup, (GDestroyNotify) gst_va_decode_picture_free); return new_picture; } static gboolean gst_va_vp9_dec_negotiate (GstVideoDecoder * decoder) { GstCapsFeatures *capsfeatures = NULL; GstVaBaseDec *base = GST_VA_BASE_DEC (decoder); GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder); GstVideoFormat format = GST_VIDEO_FORMAT_UNKNOWN; GstVp9Decoder *vp9dec = GST_VP9_DECODER (decoder); gboolean need_open; /* Ignore downstream renegotiation request. */ if (!self->need_negotiation) return TRUE; self->need_negotiation = FALSE; need_open = TRUE; /* VP9 profile entry should have the ability to handle dynamical * resolution changes. If only the resolution changes, we should not * re-create the config and context. */ if (gst_va_decoder_is_open (base->decoder)) { VAProfile cur_profile; guint cur_rtformat; gint cur_width, cur_height; if (!gst_va_decoder_get_config (base->decoder, &cur_profile, &cur_rtformat, &cur_width, &cur_height)) return FALSE; if (base->profile == cur_profile && base->rt_format == cur_rtformat) { if (!gst_va_decoder_update_frame_size (base->decoder, base->width, base->height)) return FALSE; GST_INFO_OBJECT (self, "dynamical resolution changes from %dx%d to" " %dx%d", cur_width, cur_height, base->width, base->height); need_open = FALSE; } else { if (!gst_va_decoder_close (base->decoder)) return FALSE; } } if (need_open) { if (!gst_va_decoder_open (base->decoder, base->profile, base->rt_format)) return FALSE; if (!gst_va_decoder_set_frame_size (base->decoder, base->width, base->height)) return FALSE; } if (base->output_state) gst_video_codec_state_unref (base->output_state); gst_va_base_dec_get_preferred_format_and_caps_features (base, &format, &capsfeatures); base->output_state = gst_video_decoder_set_output_state (decoder, format, base->width, base->height, vp9dec->input_state); base->output_state->caps = gst_video_info_to_caps (&base->output_state->info); if (capsfeatures) gst_caps_set_features_simple (base->output_state->caps, capsfeatures); GST_INFO_OBJECT (self, "Negotiated caps %" GST_PTR_FORMAT, base->output_state->caps); return GST_VIDEO_DECODER_CLASS (GST_VA_BASE_DEC_GET_PARENT_CLASS (decoder))->negotiate (decoder); } static void gst_va_vp9_dec_dispose (GObject * object) { gst_va_base_dec_close (GST_VIDEO_DECODER (object)); G_OBJECT_CLASS (GST_VA_BASE_DEC_GET_PARENT_CLASS (object))->dispose (object); } static void gst_va_vp9_dec_class_init (gpointer g_class, gpointer class_data) { GstCaps *src_doc_caps, *sink_doc_caps; GObjectClass *gobject_class = G_OBJECT_CLASS (g_class); GstElementClass *element_class = GST_ELEMENT_CLASS (g_class); GstVideoDecoderClass *decoder_class = GST_VIDEO_DECODER_CLASS (g_class); GstVp9DecoderClass *vp9_class = GST_VP9_DECODER_CLASS (g_class); struct CData *cdata = class_data; gchar *long_name; if (cdata->description) { long_name = g_strdup_printf ("VA-API VP9 Decoder in %s", cdata->description); } else { long_name = g_strdup ("VA-API VP9 Decoder"); } gst_element_class_set_metadata (element_class, long_name, "Codec/Decoder/Video/Hardware", "VA-API based VP9 video decoder", "Víctor Jáquez "); sink_doc_caps = gst_caps_from_string (sink_caps_str); src_doc_caps = gst_caps_from_string (src_caps_str); gst_va_base_dec_class_init (GST_VA_BASE_DEC_CLASS (g_class), VP9, cdata->render_device_path, cdata->sink_caps, cdata->src_caps, src_doc_caps, sink_doc_caps); gobject_class->dispose = gst_va_vp9_dec_dispose; decoder_class->negotiate = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_negotiate); vp9_class->new_sequence = GST_DEBUG_FUNCPTR (gst_va_vp9_new_sequence); vp9_class->new_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_new_picture); vp9_class->decode_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_decode_picture); vp9_class->end_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_end_picture); vp9_class->output_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_output_picture); vp9_class->duplicate_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_duplicate_picture); g_free (long_name); g_free (cdata->description); g_free (cdata->render_device_path); gst_caps_unref (cdata->src_caps); gst_caps_unref (cdata->sink_caps); g_free (cdata); } static void gst_va_vp9_dec_init (GTypeInstance * instance, gpointer g_class) { gst_va_base_dec_init (GST_VA_BASE_DEC (instance), GST_CAT_DEFAULT); } /* This element doesn't parse supreframes. Let's delegate it to the * parser. */ static GstCaps * _complete_sink_caps (GstCaps * sinkcaps) { gst_caps_set_simple (sinkcaps, "alignment", G_TYPE_STRING, "frame", NULL); return gst_caps_ref (sinkcaps); } static gpointer _register_debug_category (gpointer data) { GST_DEBUG_CATEGORY_INIT (gst_va_vp9dec_debug, "vavp9dec", 0, "VA VP9 decoder"); return NULL; } gboolean gst_va_vp9_dec_register (GstPlugin * plugin, GstVaDevice * device, GstCaps * sink_caps, GstCaps * src_caps, guint rank) { static GOnce debug_once = G_ONCE_INIT; GType type; GTypeInfo type_info = { .class_size = sizeof (GstVaVp9DecClass), .class_init = gst_va_vp9_dec_class_init, .instance_size = sizeof (GstVaVp9Dec), .instance_init = gst_va_vp9_dec_init, }; struct CData *cdata; gboolean ret; gchar *type_name, *feature_name; g_return_val_if_fail (GST_IS_PLUGIN (plugin), FALSE); g_return_val_if_fail (GST_IS_VA_DEVICE (device), FALSE); g_return_val_if_fail (GST_IS_CAPS (sink_caps), FALSE); g_return_val_if_fail (GST_IS_CAPS (src_caps), FALSE); cdata = g_new (struct CData, 1); cdata->description = NULL; cdata->render_device_path = g_strdup (device->render_device_path); cdata->sink_caps = _complete_sink_caps (sink_caps); cdata->src_caps = gst_caps_ref (src_caps); /* class data will be leaked if the element never gets instantiated */ GST_MINI_OBJECT_FLAG_SET (sink_caps, GST_MINI_OBJECT_FLAG_MAY_BE_LEAKED); GST_MINI_OBJECT_FLAG_SET (src_caps, GST_MINI_OBJECT_FLAG_MAY_BE_LEAKED); type_info.class_data = cdata; type_name = g_strdup ("GstVaVp9Dec"); feature_name = g_strdup ("vavp9dec"); /* The first decoder to be registered should use a constant name, * like vavp9dec, for any additional decoders, we create unique * names, using inserting the render device name. */ if (g_type_from_name (type_name)) { gchar *basename = g_path_get_basename (device->render_device_path); g_free (type_name); g_free (feature_name); type_name = g_strdup_printf ("GstVa%sVp9Dec", basename); feature_name = g_strdup_printf ("va%svp9dec", basename); cdata->description = basename; /* lower rank for non-first device */ if (rank > 0) rank--; } g_once (&debug_once, _register_debug_category, NULL); type = g_type_register_static (GST_TYPE_VP9_DECODER, type_name, &type_info, 0); ret = gst_element_register (plugin, feature_name, rank, type); g_free (type_name); g_free (feature_name); return ret; }