mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-06 09:29:42 +00:00
4a4823b972
Delay decoders downstream negotiation just before an output frame needs to be allocated. This is required, are least for H.264 and H.265 decoders, since codec_data might trigger a new sequence before finishing upstream negotiation, and sink pad caps need to set before setting source pad caps, particularly to forward HDR fields. The other decoders are changed too in order to keep the same structure among them. Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/1257>
764 lines
24 KiB
C
764 lines
24 KiB
C
/* GStreamer
|
|
* Copyright (C) 2020 Igalia, S.L.
|
|
* Author: Víctor Jáquez <vjaquez@igalia.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the0
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
/**
|
|
* SECTION:element-vavp9dec
|
|
* @title: vavp9dec
|
|
* @short_description: A VA-API based VP9 video decoder
|
|
*
|
|
* vavp9dec decodes VP9 bitstreams to VA surfaces using the
|
|
* installed and chosen [VA-API](https://01.org/linuxmedia/vaapi)
|
|
* driver.
|
|
*
|
|
* The decoding surfaces can be mapped onto main memory as video
|
|
* frames.
|
|
*
|
|
* ## Example launch line
|
|
* ```
|
|
* gst-launch-1.0 filesrc location=sample.webm ! parsebin ! vavp9dec ! autovideosink
|
|
* ```
|
|
*
|
|
* Since: 1.20
|
|
*
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include "gstvavp9dec.h"
|
|
|
|
#include "gstvabasedec.h"
|
|
|
|
GST_DEBUG_CATEGORY_STATIC (gst_va_vp9dec_debug);
|
|
#ifndef GST_DISABLE_GST_DEBUG
|
|
#define GST_CAT_DEFAULT gst_va_vp9dec_debug
|
|
#else
|
|
#define GST_CAT_DEFAULT NULL
|
|
#endif
|
|
|
|
#define GST_VA_VP9_DEC(obj) ((GstVaVp9Dec *) obj)
|
|
#define GST_VA_VP9_DEC_GET_CLASS(obj) (G_TYPE_INSTANCE_GET_CLASS ((obj), G_TYPE_FROM_INSTANCE (obj), GstVaVp9DecClass))
|
|
#define GST_VA_VP9_DEC_CLASS(klass) ((GstVaVp9DecClass *) klass)
|
|
|
|
typedef struct _GstVaVp9Dec GstVaVp9Dec;
|
|
typedef struct _GstVaVp9DecClass GstVaVp9DecClass;
|
|
|
|
struct _GstVaVp9DecClass
|
|
{
|
|
GstVaBaseDecClass parent_class;
|
|
};
|
|
|
|
struct _GstVaVp9Dec
|
|
{
|
|
GstVaBaseDec parent;
|
|
GstVp9Segmentation segmentation[GST_VP9_MAX_SEGMENTS];
|
|
};
|
|
|
|
static GstElementClass *parent_class = NULL;
|
|
|
|
/* *INDENT-OFF* */
|
|
static const gchar *src_caps_str =
|
|
GST_VIDEO_CAPS_MAKE_WITH_FEATURES (GST_CAPS_FEATURE_MEMORY_VA,
|
|
"{ NV12 }") " ;"
|
|
GST_VIDEO_CAPS_MAKE ("{ NV12 }");
|
|
/* *INDENT-ON* */
|
|
|
|
static const gchar *sink_caps_str = "video/x-vp9";
|
|
|
|
static guint
|
|
_get_rtformat (GstVaVp9Dec * self, GstVP9Profile profile,
|
|
GstVp9BitDepth bit_depth, gint subsampling_x, gint subsampling_y)
|
|
{
|
|
switch (profile) {
|
|
case GST_VP9_PROFILE_0:
|
|
return VA_RT_FORMAT_YUV420;
|
|
case GST_VP9_PROFILE_1:
|
|
if (subsampling_x == 1 && subsampling_y == 0)
|
|
return VA_RT_FORMAT_YUV422;
|
|
else if (subsampling_x == 0 && subsampling_y == 0)
|
|
return VA_RT_FORMAT_YUV444;
|
|
break;
|
|
case GST_VP9_PROFILE_2:
|
|
if (bit_depth == GST_VP9_BIT_DEPTH_10)
|
|
return VA_RT_FORMAT_YUV420_10;
|
|
else if (bit_depth == GST_VP9_BIT_DEPTH_12)
|
|
return VA_RT_FORMAT_YUV420_12;
|
|
break;
|
|
case GST_VP9_PROFILE_3:
|
|
if (subsampling_x == 1 && subsampling_y == 0) {
|
|
if (bit_depth == GST_VP9_BIT_DEPTH_10)
|
|
return VA_RT_FORMAT_YUV422_10;
|
|
else if (bit_depth == GST_VP9_BIT_DEPTH_12)
|
|
return VA_RT_FORMAT_YUV422_12;
|
|
} else if (subsampling_x == 0 && subsampling_y == 0) {
|
|
if (bit_depth == GST_VP9_BIT_DEPTH_10)
|
|
return VA_RT_FORMAT_YUV444_10;
|
|
else if (bit_depth == GST_VP9_BIT_DEPTH_12)
|
|
return VA_RT_FORMAT_YUV444_12;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
GST_ERROR_OBJECT (self, "Unsupported chroma format");
|
|
return 0;
|
|
}
|
|
|
|
static VAProfile
|
|
_get_profile (GstVaVp9Dec * self, GstVP9Profile profile)
|
|
{
|
|
switch (profile) {
|
|
case GST_VP9_PROFILE_0:
|
|
return VAProfileVP9Profile0;
|
|
case GST_VP9_PROFILE_1:
|
|
return VAProfileVP9Profile1;
|
|
case GST_VP9_PROFILE_2:
|
|
return VAProfileVP9Profile2;
|
|
case GST_VP9_PROFILE_3:
|
|
return VAProfileVP9Profile3;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
GST_ERROR_OBJECT (self, "Unsupported profile");
|
|
return VAProfileNone;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_vp9_new_sequence (GstVp9Decoder * decoder,
|
|
const GstVp9FrameHeader * frame_hdr)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder);
|
|
VAProfile profile;
|
|
gboolean negotiation_needed = FALSE;
|
|
guint rt_format;
|
|
|
|
profile = _get_profile (self, frame_hdr->profile);
|
|
if (profile == VAProfileNone)
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
|
|
if (!gst_va_decoder_has_profile (base->decoder, profile)) {
|
|
GST_ERROR_OBJECT (self, "Profile %s is not supported",
|
|
gst_va_profile_name (profile));
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
}
|
|
|
|
rt_format = _get_rtformat (self, frame_hdr->profile, frame_hdr->bit_depth,
|
|
frame_hdr->subsampling_x, frame_hdr->subsampling_y);
|
|
if (rt_format == 0)
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
|
|
if (!gst_va_decoder_config_is_equal (base->decoder, profile,
|
|
rt_format, frame_hdr->width, frame_hdr->height)) {
|
|
base->profile = profile;
|
|
base->width = frame_hdr->width;
|
|
base->height = frame_hdr->height;
|
|
base->rt_format = rt_format;
|
|
negotiation_needed = TRUE;
|
|
}
|
|
|
|
base->min_buffers = GST_VP9_REF_FRAMES;
|
|
|
|
base->need_negotiation = negotiation_needed;
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
_check_resolution_change (GstVaVp9Dec * self, GstVp9Picture * picture)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (self);
|
|
const GstVp9FrameHeader *frame_hdr = &picture->frame_hdr;
|
|
|
|
if ((base->width != frame_hdr->width) || base->height != frame_hdr->height) {
|
|
base->width = frame_hdr->width;
|
|
base->height = frame_hdr->height;
|
|
|
|
base->need_negotiation = TRUE;
|
|
if (!gst_video_decoder_negotiate (GST_VIDEO_DECODER (self))) {
|
|
GST_ERROR_OBJECT (self, "Resolution changed, but failed to"
|
|
" negotiate with downstream");
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
|
|
/* @TODO: if negotiation fails, decoder should resize output
|
|
* frame. For that we would need an auxiliar allocator, and
|
|
* later use GstVaFilter or GstVideoConverter. */
|
|
}
|
|
}
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_vp9_dec_new_picture (GstVp9Decoder * decoder,
|
|
GstVideoCodecFrame * frame, GstVp9Picture * picture)
|
|
{
|
|
GstFlowReturn ret;
|
|
GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder);
|
|
GstVaDecodePicture *pic;
|
|
GstVideoDecoder *vdec = GST_VIDEO_DECODER (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
|
|
ret = _check_resolution_change (self, picture);
|
|
if (ret != GST_FLOW_OK)
|
|
return ret;
|
|
|
|
if (base->need_negotiation) {
|
|
if (!gst_video_decoder_negotiate (vdec)) {
|
|
GST_ERROR_OBJECT (self, "Failed to negotiate with downstream");
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
}
|
|
}
|
|
|
|
ret = gst_video_decoder_allocate_output_frame (vdec, frame);
|
|
if (ret != GST_FLOW_OK)
|
|
goto error;
|
|
|
|
pic = gst_va_decode_picture_new (base->decoder, frame->output_buffer);
|
|
|
|
gst_vp9_picture_set_user_data (picture, pic,
|
|
(GDestroyNotify) gst_va_decode_picture_free);
|
|
|
|
GST_LOG_OBJECT (self, "New va decode picture %p - %#x", pic,
|
|
gst_va_decode_picture_get_surface (pic));
|
|
|
|
return GST_FLOW_OK;
|
|
|
|
error:
|
|
{
|
|
GST_WARNING_OBJECT (self, "Failed to allocated output buffer, return %s",
|
|
gst_flow_get_name (ret));
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
static inline gboolean
|
|
_fill_param (GstVp9Decoder * decoder, GstVp9Picture * picture, GstVp9Dpb * dpb)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVaDecodePicture *va_pic;
|
|
const GstVp9FrameHeader *frame_hdr = &picture->frame_hdr;
|
|
const GstVp9LoopFilterParams *lfp = &frame_hdr->loop_filter_params;
|
|
const GstVp9SegmentationParams *sp = &frame_hdr->segmentation_params;
|
|
VADecPictureParameterBufferVP9 pic_param;
|
|
guint i;
|
|
|
|
/* *INDENT-OFF* */
|
|
pic_param = (VADecPictureParameterBufferVP9) {
|
|
.frame_width = base->width,
|
|
.frame_height = base->height,
|
|
|
|
.pic_fields.bits = {
|
|
.subsampling_x = frame_hdr->subsampling_x,
|
|
.subsampling_y = frame_hdr->subsampling_x,
|
|
.frame_type = frame_hdr->frame_type,
|
|
.show_frame = frame_hdr->show_frame,
|
|
.error_resilient_mode = frame_hdr->error_resilient_mode,
|
|
.intra_only = frame_hdr->intra_only,
|
|
.allow_high_precision_mv = frame_hdr->allow_high_precision_mv,
|
|
.mcomp_filter_type = frame_hdr->interpolation_filter,
|
|
.frame_parallel_decoding_mode = frame_hdr->frame_parallel_decoding_mode,
|
|
.reset_frame_context = frame_hdr->reset_frame_context,
|
|
.refresh_frame_context = frame_hdr->refresh_frame_context,
|
|
.frame_context_idx = frame_hdr->frame_context_idx,
|
|
|
|
.segmentation_enabled = sp->segmentation_enabled,
|
|
.segmentation_temporal_update = sp->segmentation_temporal_update,
|
|
.segmentation_update_map = sp->segmentation_update_map,
|
|
|
|
.last_ref_frame =
|
|
frame_hdr->ref_frame_idx[GST_VP9_REF_FRAME_LAST - 1],
|
|
.last_ref_frame_sign_bias =
|
|
frame_hdr->ref_frame_sign_bias[GST_VP9_REF_FRAME_LAST],
|
|
.golden_ref_frame =
|
|
frame_hdr->ref_frame_idx[GST_VP9_REF_FRAME_GOLDEN - 1],
|
|
.golden_ref_frame_sign_bias =
|
|
frame_hdr->ref_frame_sign_bias[GST_VP9_REF_FRAME_GOLDEN],
|
|
.alt_ref_frame =
|
|
frame_hdr->ref_frame_idx[GST_VP9_REF_FRAME_ALTREF - 1],
|
|
.alt_ref_frame_sign_bias =
|
|
frame_hdr->ref_frame_sign_bias[GST_VP9_REF_FRAME_ALTREF],
|
|
|
|
.lossless_flag = frame_hdr->lossless_flag,
|
|
},
|
|
|
|
.filter_level = lfp->loop_filter_level,
|
|
.sharpness_level = lfp->loop_filter_sharpness,
|
|
.log2_tile_rows = frame_hdr->tile_rows_log2,
|
|
.log2_tile_columns = frame_hdr->tile_cols_log2,
|
|
|
|
.frame_header_length_in_bytes = frame_hdr->frame_header_length_in_bytes,
|
|
.first_partition_size = frame_hdr->header_size_in_bytes,
|
|
|
|
.profile = frame_hdr->profile,
|
|
.bit_depth = frame_hdr->bit_depth
|
|
};
|
|
/* *INDENT-ON* */
|
|
|
|
memcpy (pic_param.mb_segment_tree_probs, sp->segmentation_tree_probs,
|
|
sizeof (sp->segmentation_tree_probs));
|
|
|
|
if (sp->segmentation_temporal_update) {
|
|
memcpy (pic_param.segment_pred_probs, sp->segmentation_pred_prob,
|
|
sizeof (sp->segmentation_pred_prob));
|
|
} else {
|
|
memset (pic_param.segment_pred_probs, 255,
|
|
sizeof (pic_param.segment_pred_probs));
|
|
}
|
|
|
|
for (i = 0; i < GST_VP9_REF_FRAMES; i++) {
|
|
if (dpb->pic_list[i]) {
|
|
GstVaDecodePicture *va_pic =
|
|
gst_vp9_picture_get_user_data (dpb->pic_list[i]);
|
|
|
|
pic_param.reference_frames[i] =
|
|
gst_va_decode_picture_get_surface (va_pic);
|
|
} else {
|
|
pic_param.reference_frames[i] = VA_INVALID_ID;
|
|
}
|
|
}
|
|
|
|
va_pic = gst_vp9_picture_get_user_data (picture);
|
|
|
|
return gst_va_decoder_add_param_buffer (base->decoder, va_pic,
|
|
VAPictureParameterBufferType, &pic_param, sizeof (pic_param));
|
|
}
|
|
|
|
static void
|
|
_update_segmentation (GstVaVp9Dec * self, GstVp9FrameHeader * header)
|
|
{
|
|
const GstVp9LoopFilterParams *lfp = &header->loop_filter_params;
|
|
const GstVp9QuantizationParams *qp = &header->quantization_params;
|
|
const GstVp9SegmentationParams *sp = &header->segmentation_params;
|
|
guint8 n_shift = lfp->loop_filter_level >> 5;
|
|
guint i;
|
|
|
|
for (i = 0; i < GST_VP9_MAX_SEGMENTS; i++) {
|
|
gint16 luma_dc_quant_scale;
|
|
gint16 luma_ac_quant_scale;
|
|
gint16 chroma_dc_quant_scale;
|
|
gint16 chroma_ac_quant_scale;
|
|
guint8 qindex;
|
|
guint8 lvl_lookup[GST_VP9_MAX_REF_LF_DELTAS][GST_VP9_MAX_MODE_LF_DELTAS];
|
|
gint lvl_seg = lfp->loop_filter_level;
|
|
|
|
/* 8.6.1 Dequantization functions */
|
|
qindex = gst_vp9_get_qindex (sp, qp, i);
|
|
luma_dc_quant_scale =
|
|
gst_vp9_get_dc_quant (qindex, qp->delta_q_y_dc, header->bit_depth);
|
|
luma_ac_quant_scale = gst_vp9_get_ac_quant (qindex, 0, header->bit_depth);
|
|
chroma_dc_quant_scale =
|
|
gst_vp9_get_dc_quant (qindex, qp->delta_q_uv_dc, header->bit_depth);
|
|
chroma_ac_quant_scale =
|
|
gst_vp9_get_ac_quant (qindex, qp->delta_q_uv_ac, header->bit_depth);
|
|
|
|
if (!lfp->loop_filter_level) {
|
|
memset (lvl_lookup, 0, sizeof (lvl_lookup));
|
|
} else {
|
|
/* 8.8.1 Loop filter frame init process */
|
|
if (gst_vp9_seg_feature_active (sp, i, GST_VP9_SEG_LVL_ALT_L)) {
|
|
if (sp->segmentation_abs_or_delta_update) {
|
|
lvl_seg = sp->feature_data[i][GST_VP9_SEG_LVL_ALT_L];
|
|
} else {
|
|
lvl_seg += sp->feature_data[i][GST_VP9_SEG_LVL_ALT_L];
|
|
}
|
|
|
|
lvl_seg = CLAMP (lvl_seg, 0, GST_VP9_MAX_LOOP_FILTER);
|
|
}
|
|
|
|
if (!lfp->loop_filter_delta_enabled) {
|
|
memset (lvl_lookup, lvl_seg, sizeof (lvl_lookup));
|
|
} else {
|
|
guint8 ref, mode;
|
|
gint intra_lvl = lvl_seg +
|
|
(lfp->loop_filter_ref_deltas[GST_VP9_REF_FRAME_INTRA] << n_shift);
|
|
|
|
memcpy (lvl_lookup, self->segmentation[i].filter_level,
|
|
sizeof (lvl_lookup));
|
|
|
|
lvl_lookup[GST_VP9_REF_FRAME_INTRA][0] =
|
|
CLAMP (intra_lvl, 0, GST_VP9_MAX_LOOP_FILTER);
|
|
for (ref = GST_VP9_REF_FRAME_LAST; ref < GST_VP9_REF_FRAME_MAX; ref++) {
|
|
for (mode = 0; mode < GST_VP9_MAX_MODE_LF_DELTAS; mode++) {
|
|
intra_lvl = lvl_seg + (lfp->loop_filter_ref_deltas[ref] << n_shift)
|
|
+ (lfp->loop_filter_mode_deltas[mode] << n_shift);
|
|
lvl_lookup[ref][mode] =
|
|
CLAMP (intra_lvl, 0, GST_VP9_MAX_LOOP_FILTER);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/* *INDENT-OFF* */
|
|
self->segmentation[i] = (GstVp9Segmentation) {
|
|
.luma_dc_quant_scale = luma_dc_quant_scale,
|
|
.luma_ac_quant_scale = luma_ac_quant_scale,
|
|
.chroma_dc_quant_scale = chroma_dc_quant_scale,
|
|
.chroma_ac_quant_scale = chroma_ac_quant_scale,
|
|
|
|
.reference_frame_enabled = sp->feature_enabled[i][GST_VP9_SEG_LVL_REF_FRAME],
|
|
.reference_frame = sp->feature_data[i][GST_VP9_SEG_LVL_REF_FRAME],
|
|
.reference_skip = sp->feature_enabled[i][GST_VP9_SEG_SEG_LVL_SKIP],
|
|
};
|
|
/* *INDENT-ON* */
|
|
|
|
memcpy (self->segmentation[i].filter_level, lvl_lookup,
|
|
sizeof (lvl_lookup));
|
|
}
|
|
}
|
|
|
|
static inline gboolean
|
|
_fill_slice (GstVp9Decoder * decoder, GstVp9Picture * picture)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder);
|
|
GstVaDecodePicture *va_pic;
|
|
const GstVp9Segmentation *seg;
|
|
VASliceParameterBufferVP9 slice_param;
|
|
guint i;
|
|
|
|
_update_segmentation (self, &picture->frame_hdr);
|
|
|
|
/* *INDENT-OFF* */
|
|
slice_param = (VASliceParameterBufferVP9) {
|
|
.slice_data_size = picture->size,
|
|
.slice_data_offset = 0,
|
|
.slice_data_flag = VA_SLICE_DATA_FLAG_ALL,
|
|
};
|
|
/* *INDENT-ON* */
|
|
|
|
for (i = 0; i < GST_VP9_MAX_SEGMENTS; i++) {
|
|
seg = &self->segmentation[i];
|
|
|
|
/* *INDENT-OFF* */
|
|
slice_param.seg_param[i] = (VASegmentParameterVP9) {
|
|
.segment_flags.fields = {
|
|
.segment_reference_enabled = seg->reference_frame_enabled,
|
|
.segment_reference = seg->reference_frame,
|
|
.segment_reference_skipped = seg->reference_skip,
|
|
},
|
|
.luma_dc_quant_scale = seg->luma_dc_quant_scale,
|
|
.luma_ac_quant_scale = seg->luma_ac_quant_scale,
|
|
.chroma_dc_quant_scale = seg->chroma_dc_quant_scale,
|
|
.chroma_ac_quant_scale = seg->chroma_ac_quant_scale,
|
|
};
|
|
/* *INDENT-ON* */
|
|
|
|
memcpy (slice_param.seg_param[i].filter_level, seg->filter_level,
|
|
sizeof (slice_param.seg_param[i].filter_level));
|
|
}
|
|
|
|
va_pic = gst_vp9_picture_get_user_data (picture);
|
|
|
|
return gst_va_decoder_add_slice_buffer (base->decoder, va_pic, &slice_param,
|
|
sizeof (slice_param), (gpointer) picture->data, picture->size);
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_vp9_decode_picture (GstVp9Decoder * decoder, GstVp9Picture * picture,
|
|
GstVp9Dpb * dpb)
|
|
{
|
|
if (_fill_param (decoder, picture, dpb) && _fill_slice (decoder, picture))
|
|
return GST_FLOW_OK;
|
|
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_vp9_dec_end_picture (GstVp9Decoder * decoder, GstVp9Picture * picture)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVaDecodePicture *va_pic;
|
|
|
|
GST_LOG_OBJECT (base, "end picture %p", picture);
|
|
|
|
va_pic = gst_vp9_picture_get_user_data (picture);
|
|
|
|
if (!gst_va_decoder_decode (base->decoder, va_pic))
|
|
return GST_FLOW_ERROR;
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_vp9_dec_output_picture (GstVp9Decoder * decoder,
|
|
GstVideoCodecFrame * frame, GstVp9Picture * picture)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder);
|
|
|
|
GST_LOG_OBJECT (self, "Outputting picture %p", picture);
|
|
|
|
if (base->copy_frames)
|
|
gst_va_base_dec_copy_output_buffer (base, frame);
|
|
|
|
gst_vp9_picture_unref (picture);
|
|
|
|
return gst_video_decoder_finish_frame (GST_VIDEO_DECODER (self), frame);
|
|
}
|
|
|
|
static GstVp9Picture *
|
|
gst_va_vp9_dec_duplicate_picture (GstVp9Decoder * decoder,
|
|
GstVideoCodecFrame * frame, GstVp9Picture * picture)
|
|
{
|
|
GstVaDecodePicture *va_pic, *va_dup;
|
|
GstVp9Picture *new_picture;
|
|
|
|
if (_check_resolution_change (GST_VA_VP9_DEC (decoder), picture) !=
|
|
GST_FLOW_OK) {
|
|
return NULL;
|
|
}
|
|
|
|
va_pic = gst_vp9_picture_get_user_data (picture);
|
|
va_dup = gst_va_decode_picture_dup (va_pic);
|
|
|
|
new_picture = gst_vp9_picture_new ();
|
|
new_picture->frame_hdr = picture->frame_hdr;
|
|
|
|
frame->output_buffer = gst_buffer_ref (va_dup->gstbuffer);
|
|
|
|
gst_vp9_picture_set_user_data (picture, va_dup,
|
|
(GDestroyNotify) gst_va_decode_picture_free);
|
|
|
|
return new_picture;
|
|
}
|
|
|
|
static gboolean
|
|
gst_va_vp9_dec_negotiate (GstVideoDecoder * decoder)
|
|
{
|
|
GstCapsFeatures *capsfeatures = NULL;
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVaVp9Dec *self = GST_VA_VP9_DEC (decoder);
|
|
GstVideoFormat format = GST_VIDEO_FORMAT_UNKNOWN;
|
|
GstVp9Decoder *vp9dec = GST_VP9_DECODER (decoder);
|
|
gboolean need_open;
|
|
|
|
/* Ignore downstream renegotiation request. */
|
|
if (!base->need_negotiation)
|
|
return TRUE;
|
|
|
|
base->need_negotiation = FALSE;
|
|
|
|
need_open = TRUE;
|
|
/* VP9 profile entry should have the ability to handle dynamical
|
|
* resolution changes. If only the resolution changes, we should not
|
|
* re-create the config and context. */
|
|
if (gst_va_decoder_is_open (base->decoder)) {
|
|
VAProfile cur_profile;
|
|
guint cur_rtformat;
|
|
gint cur_width, cur_height;
|
|
|
|
if (!gst_va_decoder_get_config (base->decoder, &cur_profile,
|
|
&cur_rtformat, &cur_width, &cur_height))
|
|
return FALSE;
|
|
|
|
if (base->profile == cur_profile && base->rt_format == cur_rtformat) {
|
|
if (!gst_va_decoder_update_frame_size (base->decoder, base->width,
|
|
base->height))
|
|
return FALSE;
|
|
|
|
GST_INFO_OBJECT (self, "dynamical resolution changes from %dx%d to"
|
|
" %dx%d", cur_width, cur_height, base->width, base->height);
|
|
|
|
need_open = FALSE;
|
|
} else {
|
|
if (!gst_va_decoder_close (base->decoder))
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
if (need_open) {
|
|
if (!gst_va_decoder_open (base->decoder, base->profile, base->rt_format))
|
|
return FALSE;
|
|
|
|
if (!gst_va_decoder_set_frame_size (base->decoder, base->width,
|
|
base->height))
|
|
return FALSE;
|
|
}
|
|
|
|
if (base->output_state)
|
|
gst_video_codec_state_unref (base->output_state);
|
|
|
|
gst_va_base_dec_get_preferred_format_and_caps_features (base, &format,
|
|
&capsfeatures);
|
|
|
|
base->output_state =
|
|
gst_video_decoder_set_output_state (decoder, format,
|
|
base->width, base->height, vp9dec->input_state);
|
|
|
|
base->output_state->caps = gst_video_info_to_caps (&base->output_state->info);
|
|
if (capsfeatures)
|
|
gst_caps_set_features_simple (base->output_state->caps, capsfeatures);
|
|
|
|
GST_INFO_OBJECT (self, "Negotiated caps %" GST_PTR_FORMAT,
|
|
base->output_state->caps);
|
|
|
|
return GST_VIDEO_DECODER_CLASS (parent_class)->negotiate (decoder);
|
|
}
|
|
|
|
static void
|
|
gst_va_vp9_dec_dispose (GObject * object)
|
|
{
|
|
gst_va_base_dec_close (GST_VIDEO_DECODER (object));
|
|
G_OBJECT_CLASS (parent_class)->dispose (object);
|
|
}
|
|
|
|
static void
|
|
gst_va_vp9_dec_class_init (gpointer g_class, gpointer class_data)
|
|
{
|
|
GstCaps *src_doc_caps, *sink_doc_caps;
|
|
GObjectClass *gobject_class = G_OBJECT_CLASS (g_class);
|
|
GstElementClass *element_class = GST_ELEMENT_CLASS (g_class);
|
|
GstVideoDecoderClass *decoder_class = GST_VIDEO_DECODER_CLASS (g_class);
|
|
GstVp9DecoderClass *vp9_class = GST_VP9_DECODER_CLASS (g_class);
|
|
struct CData *cdata = class_data;
|
|
gchar *long_name;
|
|
|
|
if (cdata->description) {
|
|
long_name = g_strdup_printf ("VA-API VP9 Decoder in %s",
|
|
cdata->description);
|
|
} else {
|
|
long_name = g_strdup ("VA-API VP9 Decoder");
|
|
}
|
|
|
|
gst_element_class_set_metadata (element_class, long_name,
|
|
"Codec/Decoder/Video/Hardware", "VA-API based VP9 video decoder",
|
|
"Víctor Jáquez <vjaquez@igalia.com>");
|
|
|
|
sink_doc_caps = gst_caps_from_string (sink_caps_str);
|
|
src_doc_caps = gst_caps_from_string (src_caps_str);
|
|
|
|
parent_class = g_type_class_peek_parent (g_class);
|
|
|
|
gst_va_base_dec_class_init (GST_VA_BASE_DEC_CLASS (g_class), VP9,
|
|
cdata->render_device_path, cdata->sink_caps, cdata->src_caps,
|
|
src_doc_caps, sink_doc_caps);
|
|
|
|
gobject_class->dispose = gst_va_vp9_dec_dispose;
|
|
|
|
decoder_class->negotiate = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_negotiate);
|
|
|
|
vp9_class->new_sequence = GST_DEBUG_FUNCPTR (gst_va_vp9_new_sequence);
|
|
vp9_class->new_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_new_picture);
|
|
vp9_class->decode_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_decode_picture);
|
|
vp9_class->end_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_end_picture);
|
|
vp9_class->output_picture = GST_DEBUG_FUNCPTR (gst_va_vp9_dec_output_picture);
|
|
vp9_class->duplicate_picture =
|
|
GST_DEBUG_FUNCPTR (gst_va_vp9_dec_duplicate_picture);
|
|
|
|
g_free (long_name);
|
|
g_free (cdata->description);
|
|
g_free (cdata->render_device_path);
|
|
gst_caps_unref (cdata->src_caps);
|
|
gst_caps_unref (cdata->sink_caps);
|
|
g_free (cdata);
|
|
}
|
|
|
|
static void
|
|
gst_va_vp9_dec_init (GTypeInstance * instance, gpointer g_class)
|
|
{
|
|
gst_va_base_dec_init (GST_VA_BASE_DEC (instance), GST_CAT_DEFAULT);
|
|
}
|
|
|
|
/* This element doesn't parse supreframes. Let's delegate it to the
|
|
* parser. */
|
|
static GstCaps *
|
|
_complete_sink_caps (GstCaps * sinkcaps)
|
|
{
|
|
gst_caps_set_simple (sinkcaps, "alignment", G_TYPE_STRING, "frame", NULL);
|
|
return gst_caps_ref (sinkcaps);
|
|
}
|
|
|
|
static gpointer
|
|
_register_debug_category (gpointer data)
|
|
{
|
|
GST_DEBUG_CATEGORY_INIT (gst_va_vp9dec_debug, "vavp9dec", 0,
|
|
"VA VP9 decoder");
|
|
|
|
return NULL;
|
|
}
|
|
|
|
gboolean
|
|
gst_va_vp9_dec_register (GstPlugin * plugin, GstVaDevice * device,
|
|
GstCaps * sink_caps, GstCaps * src_caps, guint rank)
|
|
{
|
|
static GOnce debug_once = G_ONCE_INIT;
|
|
GType type;
|
|
GTypeInfo type_info = {
|
|
.class_size = sizeof (GstVaVp9DecClass),
|
|
.class_init = gst_va_vp9_dec_class_init,
|
|
.instance_size = sizeof (GstVaVp9Dec),
|
|
.instance_init = gst_va_vp9_dec_init,
|
|
};
|
|
struct CData *cdata;
|
|
gboolean ret;
|
|
gchar *type_name, *feature_name;
|
|
|
|
g_return_val_if_fail (GST_IS_PLUGIN (plugin), FALSE);
|
|
g_return_val_if_fail (GST_IS_VA_DEVICE (device), FALSE);
|
|
g_return_val_if_fail (GST_IS_CAPS (sink_caps), FALSE);
|
|
g_return_val_if_fail (GST_IS_CAPS (src_caps), FALSE);
|
|
|
|
cdata = g_new (struct CData, 1);
|
|
cdata->description = NULL;
|
|
cdata->render_device_path = g_strdup (device->render_device_path);
|
|
cdata->sink_caps = _complete_sink_caps (sink_caps);
|
|
cdata->src_caps = gst_caps_ref (src_caps);
|
|
|
|
/* class data will be leaked if the element never gets instantiated */
|
|
GST_MINI_OBJECT_FLAG_SET (sink_caps, GST_MINI_OBJECT_FLAG_MAY_BE_LEAKED);
|
|
GST_MINI_OBJECT_FLAG_SET (src_caps, GST_MINI_OBJECT_FLAG_MAY_BE_LEAKED);
|
|
|
|
type_info.class_data = cdata;
|
|
|
|
type_name = g_strdup ("GstVaVp9Dec");
|
|
feature_name = g_strdup ("vavp9dec");
|
|
|
|
/* The first decoder to be registered should use a constant name,
|
|
* like vavp9dec, for any additional decoders, we create unique
|
|
* names, using inserting the render device name. */
|
|
if (g_type_from_name (type_name)) {
|
|
gchar *basename = g_path_get_basename (device->render_device_path);
|
|
g_free (type_name);
|
|
g_free (feature_name);
|
|
type_name = g_strdup_printf ("GstVa%sVp9Dec", basename);
|
|
feature_name = g_strdup_printf ("va%svp9dec", basename);
|
|
cdata->description = basename;
|
|
|
|
/* lower rank for non-first device */
|
|
if (rank > 0)
|
|
rank--;
|
|
}
|
|
|
|
g_once (&debug_once, _register_debug_category, NULL);
|
|
|
|
type = g_type_register_static (GST_TYPE_VP9_DECODER,
|
|
type_name, &type_info, 0);
|
|
|
|
ret = gst_element_register (plugin, feature_name, rank, type);
|
|
|
|
g_free (type_name);
|
|
g_free (feature_name);
|
|
|
|
return ret;
|
|
}
|