mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-23 10:11:08 +00:00
4a22cc8fb3
VA drivers allocate surfaces given their properties, so there's no need to provide a buffer size to the VA pool. Though, the buffer size is provided by the driver, or the canonical size is used for single planed surfaces. This patch removes the need to provide a size for the function gst_va_pool_new_with_config() and adds a helper method to retrieve the surface size, gst_va_pool_get_buffer_size(). Also change the callers accordingly. Changes for custom VA pool creation will be addressed in the following commits. Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/5805>
1131 lines
37 KiB
C
1131 lines
37 KiB
C
/* GStreamer
|
|
* Copyright (C) 2020 Intel Corporation
|
|
* Author: He Junyan <junyan.he@intel.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the0
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
/**
|
|
* SECTION:element-vaav1dec
|
|
* @title: vaav1dec
|
|
* @short_description: A VA-API based AV1 video decoder
|
|
*
|
|
* vaav1dec decodes AV1 bitstreams to VA surfaces using the
|
|
* installed and chosen [VA-API](https://01.org/linuxmedia/vaapi)
|
|
* driver.
|
|
*
|
|
* The decoding surfaces can be mapped onto main memory as video
|
|
* frames.
|
|
*
|
|
* ## Example launch line
|
|
* ```
|
|
* gst-launch-1.0 filesrc location=sample.av1 ! ivfparse ! av1parse ! vaav1dec ! autovideosink
|
|
* ```
|
|
*
|
|
* Since: 1.20
|
|
*
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include <gst/va/gstva.h>
|
|
#include <gst/va/vasurfaceimage.h>
|
|
|
|
#include "gstvaav1dec.h"
|
|
#include "gstvabasedec.h"
|
|
|
|
GST_DEBUG_CATEGORY_STATIC (gst_va_av1dec_debug);
|
|
#ifndef GST_DISABLE_GST_DEBUG
|
|
#define GST_CAT_DEFAULT gst_va_av1dec_debug
|
|
#else
|
|
#define GST_CAT_DEFAULT NULL
|
|
#endif
|
|
|
|
#define GST_VA_AV1_DEC(obj) ((GstVaAV1Dec *) obj)
|
|
#define GST_VA_AV1_DEC_GET_CLASS(obj) (G_TYPE_INSTANCE_GET_CLASS ((obj), G_TYPE_FROM_INSTANCE (obj), GstVaAV1DecClass))
|
|
#define GST_VA_AV1_DEC_CLASS(klass) ((GstVaAV1DecClass *) klass)
|
|
|
|
typedef struct _GstVaAV1Dec GstVaAV1Dec;
|
|
typedef struct _GstVaAV1DecClass GstVaAV1DecClass;
|
|
|
|
struct _GstVaAV1DecClass
|
|
{
|
|
GstVaBaseDecClass parent_class;
|
|
};
|
|
|
|
struct _GstVaAV1Dec
|
|
{
|
|
GstVaBaseDec parent;
|
|
|
|
GstAV1SequenceHeaderOBU seq;
|
|
GstVideoFormat preferred_format;
|
|
/* Used for layers not output. */
|
|
GstBufferPool *internal_pool;
|
|
};
|
|
|
|
static GstElementClass *parent_class = NULL;
|
|
|
|
/* *INDENT-OFF* */
|
|
static const gchar *src_caps_str =
|
|
GST_VIDEO_CAPS_MAKE_WITH_FEATURES (GST_CAPS_FEATURE_MEMORY_VA,
|
|
"{ NV12, P010_10LE }") " ;"
|
|
GST_VIDEO_CAPS_MAKE ("{ NV12, P010_10LE }");
|
|
/* *INDENT-ON* */
|
|
|
|
static const gchar *sink_caps_str = "video/x-av1";
|
|
|
|
static gboolean
|
|
gst_va_av1_dec_negotiate (GstVideoDecoder * decoder)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
|
|
/* Ignore downstream renegotiation request. */
|
|
if (!base->need_negotiation)
|
|
return TRUE;
|
|
|
|
base->need_negotiation = FALSE;
|
|
|
|
/* Do not re-create the context if only the frame size changes */
|
|
if (!gst_va_decoder_config_is_equal (base->decoder, base->profile,
|
|
base->rt_format, base->width, base->height)) {
|
|
if (gst_va_decoder_is_open (base->decoder)
|
|
&& !gst_va_decoder_close (base->decoder))
|
|
return FALSE;
|
|
|
|
if (!gst_va_decoder_open (base->decoder, base->profile, base->rt_format))
|
|
return FALSE;
|
|
|
|
if (!gst_va_decoder_set_frame_size (base->decoder, base->width,
|
|
base->height))
|
|
return FALSE;
|
|
}
|
|
|
|
if (!gst_va_base_dec_set_output_state (base))
|
|
return FALSE;
|
|
|
|
if (self->preferred_format != GST_VIDEO_FORMAT_UNKNOWN &&
|
|
self->preferred_format !=
|
|
GST_VIDEO_INFO_FORMAT (&base->output_state->info)) {
|
|
GST_WARNING_OBJECT (self, "The preferred_format is different from"
|
|
" the last result");
|
|
return FALSE;
|
|
}
|
|
self->preferred_format = GST_VIDEO_INFO_FORMAT (&base->output_state->info);
|
|
|
|
return GST_VIDEO_DECODER_CLASS (parent_class)->negotiate (decoder);
|
|
}
|
|
|
|
static GstCaps *
|
|
_complete_sink_caps (GstCaps * sinkcaps)
|
|
{
|
|
GstCaps *caps = gst_caps_copy (sinkcaps);
|
|
GValue val = G_VALUE_INIT;
|
|
|
|
g_value_init (&val, G_TYPE_STRING);
|
|
g_value_set_string (&val, "frame");
|
|
gst_caps_set_value (caps, "alignment", &val);
|
|
g_value_unset (&val);
|
|
|
|
return caps;
|
|
}
|
|
|
|
static VAProfile
|
|
_get_profile (GstVaAV1Dec * self, const GstAV1SequenceHeaderOBU * seq_hdr)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (self);
|
|
VAProfile profile = VAProfileNone;
|
|
|
|
switch (seq_hdr->seq_profile) {
|
|
case GST_AV1_PROFILE_0:
|
|
profile = VAProfileAV1Profile0;
|
|
break;
|
|
case GST_AV1_PROFILE_1:
|
|
profile = VAProfileAV1Profile1;
|
|
break;
|
|
default:
|
|
GST_ERROR_OBJECT (self, "Unsupported av1 profile value %d",
|
|
seq_hdr->seq_profile);
|
|
return VAProfileNone;
|
|
}
|
|
|
|
if (!gst_va_decoder_has_profile (base->decoder, profile)) {
|
|
GST_ERROR_OBJECT (self, "Profile %s is not supported by HW",
|
|
gst_va_profile_name (profile));
|
|
return VAProfileNone;
|
|
}
|
|
|
|
return profile;
|
|
}
|
|
|
|
static guint
|
|
_get_rtformat (GstVaAV1Dec * self, VAProfile profile,
|
|
const GstAV1SequenceHeaderOBU * seq_header)
|
|
{
|
|
/* 6.4.1:
|
|
seq_profile Bit depth Monochrome support Chroma subsampling
|
|
0 8 or 10 Yes YUV 4:2:0
|
|
1 8 or 10 No YUV 4:4:4
|
|
2 8 or 10 Yes YUV 4:2:2
|
|
2 12 Yes YUV 4:2:0,YUV 4:2:2,YUV 4:4:4
|
|
*/
|
|
|
|
/* TODO: consider Monochrome case. Just return 4:2:0 for Monochrome now. */
|
|
switch (profile) {
|
|
case VAProfileAV1Profile0:
|
|
if (seq_header->bit_depth == 8) {
|
|
return VA_RT_FORMAT_YUV420;
|
|
} else if (seq_header->bit_depth == 10) {
|
|
return VA_RT_FORMAT_YUV420_10;
|
|
}
|
|
break;
|
|
case VAProfileAV1Profile1:
|
|
if (seq_header->bit_depth == 8) {
|
|
return VA_RT_FORMAT_YUV444;
|
|
} else if (seq_header->bit_depth == 10) {
|
|
return VA_RT_FORMAT_YUV444_10;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
GST_ERROR_OBJECT (self, "Fail to find rtformat for profile:%s, bit_depth:%d",
|
|
gst_va_profile_name (profile), seq_header->bit_depth);
|
|
return 0;
|
|
}
|
|
|
|
static GstCaps *
|
|
gst_va_av1_dec_getcaps (GstVideoDecoder * decoder, GstCaps * filter)
|
|
{
|
|
GstCaps *sinkcaps, *caps = NULL, *tmp;
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
|
|
if (base->decoder)
|
|
caps = gst_va_decoder_get_sinkpad_caps (base->decoder);
|
|
|
|
if (caps) {
|
|
sinkcaps = _complete_sink_caps (caps);
|
|
gst_caps_unref (caps);
|
|
if (filter) {
|
|
tmp = gst_caps_intersect_full (filter, sinkcaps,
|
|
GST_CAPS_INTERSECT_FIRST);
|
|
gst_caps_unref (sinkcaps);
|
|
caps = tmp;
|
|
} else {
|
|
caps = sinkcaps;
|
|
}
|
|
GST_LOG_OBJECT (base, "Returning caps %" GST_PTR_FORMAT, caps);
|
|
} else if (!caps) {
|
|
caps = gst_video_decoder_proxy_getcaps (decoder, NULL, filter);
|
|
}
|
|
|
|
return caps;
|
|
}
|
|
|
|
static void
|
|
_clear_internal_pool (GstVaAV1Dec * self)
|
|
{
|
|
if (self->internal_pool)
|
|
gst_buffer_pool_set_active (self->internal_pool, FALSE);
|
|
|
|
gst_clear_object (&self->internal_pool);
|
|
}
|
|
|
|
static GstBufferPool *
|
|
_create_internal_pool (GstVaAV1Dec * self, gint width, gint height)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (self);
|
|
GstVideoInfo info;
|
|
GArray *surface_formats;
|
|
GstAllocator *allocator;
|
|
GstCaps *caps = NULL;
|
|
GstBufferPool *pool;
|
|
guint usage_hint;
|
|
GstAllocationParams params = { 0, };
|
|
|
|
gst_allocation_params_init (¶ms);
|
|
|
|
/* We may come here before the negotiation, make sure all pools
|
|
use the same video format. */
|
|
if (self->preferred_format == GST_VIDEO_FORMAT_UNKNOWN) {
|
|
GstVideoFormat format;
|
|
|
|
gst_va_base_dec_get_preferred_format_and_caps_features (base,
|
|
&format, NULL, NULL);
|
|
if (format == GST_VIDEO_FORMAT_UNKNOWN) {
|
|
GST_WARNING_OBJECT (self, "Failed to get format for internal pool");
|
|
return NULL;
|
|
}
|
|
|
|
self->preferred_format = format;
|
|
}
|
|
|
|
gst_video_info_set_format (&info, self->preferred_format, width, height);
|
|
|
|
caps = gst_video_info_to_caps (&info);
|
|
if (!caps) {
|
|
GST_WARNING_OBJECT (self, "Failed to create caps for internal pool");
|
|
return NULL;
|
|
}
|
|
|
|
gst_caps_set_features_simple (caps,
|
|
gst_caps_features_from_string (GST_CAPS_FEATURE_MEMORY_VA));
|
|
|
|
surface_formats = gst_va_decoder_get_surface_formats (base->decoder);
|
|
allocator = gst_va_allocator_new (base->display, surface_formats);
|
|
|
|
usage_hint = va_get_surface_usage_hint (base->display,
|
|
VAEntrypointVLD, GST_PAD_SRC, FALSE);
|
|
|
|
pool = gst_va_pool_new_with_config (caps, 1, 0, usage_hint,
|
|
GST_VA_FEATURE_AUTO, allocator, ¶ms);
|
|
|
|
gst_clear_caps (&caps);
|
|
gst_object_unref (allocator);
|
|
|
|
if (!pool) {
|
|
GST_WARNING_OBJECT (self, "Failed to create internal pool");
|
|
return NULL;
|
|
}
|
|
|
|
if (!gst_buffer_pool_set_active (pool, TRUE)) {
|
|
GST_WARNING_OBJECT (self, "Failed to activate internal pool");
|
|
gst_object_unref (pool);
|
|
return NULL;
|
|
}
|
|
|
|
return pool;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_av1_dec_new_sequence (GstAV1Decoder * decoder,
|
|
const GstAV1SequenceHeaderOBU * seq_hdr, gint max_dpb_size)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVideoInfo *info = &base->output_info;
|
|
VAProfile profile;
|
|
guint rt_format;
|
|
gint width, height;
|
|
|
|
GST_LOG_OBJECT (self, "new sequence");
|
|
|
|
profile = _get_profile (self, seq_hdr);
|
|
if (profile == VAProfileNone)
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
|
|
rt_format = _get_rtformat (self, profile, seq_hdr);
|
|
if (!rt_format)
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
|
|
self->seq = *seq_hdr;
|
|
|
|
width = seq_hdr->max_frame_width_minus_1 + 1;
|
|
height = seq_hdr->max_frame_height_minus_1 + 1;
|
|
|
|
if (!gst_va_decoder_config_is_equal (base->decoder, profile,
|
|
rt_format, width, height)) {
|
|
_clear_internal_pool (self);
|
|
self->preferred_format = GST_VIDEO_FORMAT_UNKNOWN;
|
|
|
|
base->profile = profile;
|
|
base->rt_format = rt_format;
|
|
GST_VIDEO_INFO_WIDTH (info) = base->width = width;
|
|
GST_VIDEO_INFO_HEIGHT (info) = base->height = height;
|
|
base->need_negotiation = TRUE;
|
|
base->min_buffers = 7 + 4; /* dpb size + scratch surfaces */
|
|
base->need_valign = FALSE;
|
|
}
|
|
|
|
g_clear_pointer (&base->input_state, gst_video_codec_state_unref);
|
|
base->input_state = gst_video_codec_state_ref (decoder->input_state);
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static inline GstFlowReturn
|
|
_acquire_internal_buffer (GstVaAV1Dec * self, GstVideoCodecFrame * frame)
|
|
{
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (self);
|
|
GstFlowReturn ret;
|
|
|
|
if (!self->internal_pool) {
|
|
self->internal_pool =
|
|
_create_internal_pool (self, base->width, base->height);
|
|
if (!self->internal_pool)
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
if (base->need_negotiation) {
|
|
if (!gst_video_decoder_negotiate (GST_VIDEO_DECODER (self)))
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
}
|
|
|
|
ret = gst_buffer_pool_acquire_buffer (self->internal_pool,
|
|
&frame->output_buffer, NULL);
|
|
if (ret != GST_FLOW_OK) {
|
|
GST_WARNING_OBJECT (self,
|
|
"Failed to allocated output buffer from internal pool, return %s",
|
|
gst_flow_get_name (ret));
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_av1_dec_new_picture (GstAV1Decoder * decoder,
|
|
GstVideoCodecFrame * frame, GstAV1Picture * picture)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstAV1FrameHeaderOBU *frame_hdr = &picture->frame_hdr;
|
|
GstVaDecodePicture *pic;
|
|
GstVideoInfo *info = &base->output_info;
|
|
GstFlowReturn ret;
|
|
|
|
/* Only output the highest spatial layer. For non output pictures,
|
|
we just use internal pool, then no negotiation needed. */
|
|
if (picture->spatial_id < decoder->highest_spatial_layer) {
|
|
ret = _acquire_internal_buffer (self, frame);
|
|
if (ret != GST_FLOW_OK)
|
|
return ret;
|
|
} else {
|
|
if (frame_hdr->upscaled_width != GST_VIDEO_INFO_WIDTH (info)
|
|
|| frame_hdr->frame_height != GST_VIDEO_INFO_HEIGHT (info)) {
|
|
GST_VIDEO_INFO_WIDTH (info) = frame_hdr->upscaled_width;
|
|
GST_VIDEO_INFO_HEIGHT (info) = frame_hdr->frame_height;
|
|
|
|
if (GST_VIDEO_INFO_WIDTH (info) < base->width
|
|
|| GST_VIDEO_INFO_HEIGHT (info) < base->height) {
|
|
base->need_valign = TRUE;
|
|
/* *INDENT-OFF* */
|
|
base->valign = (GstVideoAlignment) {
|
|
.padding_bottom = base->height - GST_VIDEO_INFO_HEIGHT (info),
|
|
.padding_right = base->width - GST_VIDEO_INFO_WIDTH (info),
|
|
};
|
|
/* *INDENT-ON* */
|
|
}
|
|
|
|
base->need_negotiation = TRUE;
|
|
}
|
|
|
|
ret = gst_va_base_dec_prepare_output_frame (base, frame);
|
|
if (ret != GST_FLOW_OK) {
|
|
GST_WARNING_OBJECT (self, "Failed to allocated output buffer, return %s",
|
|
gst_flow_get_name (ret));
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if (picture->apply_grain) {
|
|
if (!gst_va_buffer_create_aux_surface (frame->output_buffer)) {
|
|
GST_WARNING_OBJECT (self,
|
|
"Failed to allocated aux surface for buffer %p",
|
|
frame->output_buffer);
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
}
|
|
|
|
pic = gst_va_decode_picture_new (base->decoder, frame->output_buffer);
|
|
|
|
gst_av1_picture_set_user_data (picture, pic,
|
|
(GDestroyNotify) gst_va_decode_picture_free);
|
|
|
|
if (picture->apply_grain) {
|
|
GST_LOG_OBJECT (self, "New va decode picture %p - %#x(aux: %#x)", pic,
|
|
gst_va_decode_picture_get_surface (pic),
|
|
gst_va_decode_picture_get_aux_surface (pic));
|
|
} else {
|
|
GST_LOG_OBJECT (self, "New va decode picture %p - %#x", pic,
|
|
gst_va_decode_picture_get_surface (pic));
|
|
}
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstAV1Picture *
|
|
gst_va_av1_dec_duplicate_picture (GstAV1Decoder * decoder,
|
|
GstVideoCodecFrame * frame, GstAV1Picture * picture)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVaDecodePicture *pic;
|
|
GstVaDecodePicture *new_pic;
|
|
GstAV1Picture *new_picture;
|
|
|
|
pic = gst_av1_picture_get_user_data (picture);
|
|
if (!pic) {
|
|
GST_ERROR_OBJECT (self, "Parent picture does not have a va picture");
|
|
return NULL;
|
|
}
|
|
|
|
new_picture = gst_av1_picture_new ();
|
|
g_assert (pic->gstbuffer);
|
|
new_pic = gst_va_decode_picture_new (base->decoder, pic->gstbuffer);
|
|
|
|
GST_LOG_OBJECT (self, "Duplicate output with buffer %" GST_PTR_FORMAT
|
|
" (surface %#x)", pic, gst_va_decode_picture_get_surface (pic));
|
|
|
|
gst_av1_picture_set_user_data (new_picture, new_pic,
|
|
(GDestroyNotify) gst_va_decode_picture_free);
|
|
|
|
return new_picture;
|
|
}
|
|
|
|
static void
|
|
_setup_segment_info (VADecPictureParameterBufferAV1 * pic_param,
|
|
GstAV1FrameHeaderOBU * frame_header)
|
|
{
|
|
guint i, j;
|
|
uint8_t feature_mask;
|
|
|
|
for (i = 0; i < GST_AV1_MAX_SEGMENTS; i++)
|
|
for (j = 0; j < GST_AV1_SEG_LVL_MAX; j++)
|
|
pic_param->seg_info.feature_data[i][j] =
|
|
frame_header->segmentation_params.feature_data[i][j];
|
|
|
|
for (i = 0; i < GST_AV1_MAX_SEGMENTS; i++) {
|
|
feature_mask = 0;
|
|
for (j = 0; j < GST_AV1_SEG_LVL_MAX; j++) {
|
|
if (frame_header->segmentation_params.feature_enabled[i][j])
|
|
feature_mask |= 1 << j;
|
|
}
|
|
pic_param->seg_info.feature_mask[i] = feature_mask;
|
|
}
|
|
}
|
|
|
|
static void
|
|
_setup_film_grain_info (VADecPictureParameterBufferAV1 * pic_param,
|
|
GstAV1FrameHeaderOBU * frame_header)
|
|
{
|
|
guint i;
|
|
|
|
if (!frame_header->film_grain_params.apply_grain)
|
|
return;
|
|
|
|
pic_param->film_grain_info.num_y_points =
|
|
frame_header->film_grain_params.num_y_points;
|
|
for (i = 0; i < frame_header->film_grain_params.num_y_points; i++) {
|
|
pic_param->film_grain_info.point_y_value[i] =
|
|
frame_header->film_grain_params.point_y_value[i];
|
|
pic_param->film_grain_info.point_y_scaling[i] =
|
|
frame_header->film_grain_params.point_y_scaling[i];
|
|
}
|
|
|
|
pic_param->film_grain_info.num_cb_points =
|
|
frame_header->film_grain_params.num_cb_points;
|
|
for (i = 0; i < frame_header->film_grain_params.num_cb_points; i++) {
|
|
pic_param->film_grain_info.point_cb_value[i] =
|
|
frame_header->film_grain_params.point_cb_value[i];
|
|
pic_param->film_grain_info.point_cb_scaling[i] =
|
|
frame_header->film_grain_params.point_cb_scaling[i];
|
|
}
|
|
|
|
pic_param->film_grain_info.num_cr_points =
|
|
frame_header->film_grain_params.num_cr_points;
|
|
for (i = 0; i < frame_header->film_grain_params.num_cr_points; i++) {
|
|
pic_param->film_grain_info.point_cr_value[i] =
|
|
frame_header->film_grain_params.point_cr_value[i];
|
|
pic_param->film_grain_info.point_cr_scaling[i] =
|
|
frame_header->film_grain_params.point_cr_scaling[i];
|
|
}
|
|
|
|
|
|
if (pic_param->film_grain_info.num_y_points) {
|
|
for (i = 0; i < 24; i++) {
|
|
pic_param->film_grain_info.ar_coeffs_y[i] =
|
|
frame_header->film_grain_params.ar_coeffs_y_plus_128[i] - 128;
|
|
}
|
|
}
|
|
if (frame_header->film_grain_params.chroma_scaling_from_luma
|
|
|| pic_param->film_grain_info.num_cb_points) {
|
|
for (i = 0; i < GST_AV1_MAX_NUM_POS_LUMA; i++) {
|
|
pic_param->film_grain_info.ar_coeffs_cb[i] =
|
|
frame_header->film_grain_params.ar_coeffs_cb_plus_128[i] - 128;
|
|
}
|
|
}
|
|
if (frame_header->film_grain_params.chroma_scaling_from_luma
|
|
|| pic_param->film_grain_info.num_cr_points) {
|
|
for (i = 0; i < GST_AV1_MAX_NUM_POS_LUMA; i++) {
|
|
pic_param->film_grain_info.ar_coeffs_cr[i] =
|
|
frame_header->film_grain_params.ar_coeffs_cr_plus_128[i] - 128;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
_setup_loop_filter_info (VADecPictureParameterBufferAV1 * pic_param,
|
|
GstAV1FrameHeaderOBU * frame_header)
|
|
{
|
|
guint i;
|
|
|
|
pic_param->filter_level[0] =
|
|
frame_header->loop_filter_params.loop_filter_level[0];
|
|
pic_param->filter_level[1] =
|
|
frame_header->loop_filter_params.loop_filter_level[1];
|
|
pic_param->filter_level_u =
|
|
frame_header->loop_filter_params.loop_filter_level[2];
|
|
pic_param->filter_level_v =
|
|
frame_header->loop_filter_params.loop_filter_level[3];
|
|
|
|
for (i = 0; i < GST_AV1_TOTAL_REFS_PER_FRAME; i++)
|
|
pic_param->ref_deltas[i] =
|
|
frame_header->loop_filter_params.loop_filter_ref_deltas[i];
|
|
for (i = 0; i < 2; i++)
|
|
pic_param->mode_deltas[i] =
|
|
frame_header->loop_filter_params.loop_filter_mode_deltas[i];
|
|
}
|
|
|
|
static void
|
|
_setup_quantization_info (VADecPictureParameterBufferAV1 * pic_param,
|
|
GstAV1FrameHeaderOBU * frame_header)
|
|
{
|
|
pic_param->qmatrix_fields.bits.using_qmatrix =
|
|
frame_header->quantization_params.using_qmatrix;
|
|
if (frame_header->quantization_params.using_qmatrix) {
|
|
pic_param->qmatrix_fields.bits.qm_y =
|
|
frame_header->quantization_params.qm_y;
|
|
pic_param->qmatrix_fields.bits.qm_u =
|
|
frame_header->quantization_params.qm_u;
|
|
pic_param->qmatrix_fields.bits.qm_v =
|
|
frame_header->quantization_params.qm_v;
|
|
} else {
|
|
pic_param->qmatrix_fields.bits.qm_y = 0;
|
|
pic_param->qmatrix_fields.bits.qm_u = 0;
|
|
pic_param->qmatrix_fields.bits.qm_v = 0;
|
|
}
|
|
}
|
|
|
|
static void
|
|
_setup_cdef_info (VADecPictureParameterBufferAV1 * pic_param,
|
|
GstAV1FrameHeaderOBU * frame_header, guint8 num_planes)
|
|
{
|
|
guint8 sec_strength;
|
|
guint i;
|
|
|
|
pic_param->cdef_damping_minus_3 = frame_header->cdef_params.cdef_damping - 3;
|
|
pic_param->cdef_bits = frame_header->cdef_params.cdef_bits;
|
|
for (i = 0; i < GST_AV1_CDEF_MAX; i++) {
|
|
sec_strength = frame_header->cdef_params.cdef_y_sec_strength[i];
|
|
g_assert (sec_strength <= 4);
|
|
/* may need to minus 1 in order to merge with primary value. */
|
|
if (sec_strength == 4)
|
|
sec_strength--;
|
|
|
|
pic_param->cdef_y_strengths[i] =
|
|
((frame_header->cdef_params.cdef_y_pri_strength[i] & 0xf) << 2) |
|
|
(sec_strength & 0x03);
|
|
}
|
|
if (num_planes > 1) {
|
|
for (i = 0; i < GST_AV1_CDEF_MAX; i++) {
|
|
sec_strength = frame_header->cdef_params.cdef_uv_sec_strength[i];
|
|
g_assert (sec_strength <= 4);
|
|
/* may need to minus 1 in order to merge with primary value. */
|
|
if (sec_strength == 4)
|
|
sec_strength--;
|
|
|
|
pic_param->cdef_uv_strengths[i] =
|
|
((frame_header->cdef_params.cdef_uv_pri_strength[i] & 0xf) << 2) |
|
|
(sec_strength & 0x03);
|
|
}
|
|
} else {
|
|
for (i = 0; i < GST_AV1_CDEF_MAX; i++) {
|
|
pic_param->cdef_uv_strengths[i] = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
_setup_global_motion_info (VADecPictureParameterBufferAV1 * pic_param,
|
|
GstAV1FrameHeaderOBU * frame_header)
|
|
{
|
|
guint i, j;
|
|
|
|
for (i = 0; i < 7; i++) {
|
|
/* assuming VAAV1TransformationType and GstAV1WarpModelType are
|
|
* equivalent */
|
|
pic_param->wm[i].wmtype = (VAAV1TransformationType)
|
|
frame_header->global_motion_params.gm_type[GST_AV1_REF_LAST_FRAME + i];
|
|
|
|
for (j = 0; j < 6; j++)
|
|
pic_param->wm[i].wmmat[j] =
|
|
frame_header->global_motion_params.gm_params
|
|
[GST_AV1_REF_LAST_FRAME + i][j];
|
|
|
|
pic_param->wm[i].wmmat[6] = 0;
|
|
pic_param->wm[i].wmmat[7] = 0;
|
|
|
|
pic_param->wm[i].invalid =
|
|
frame_header->global_motion_params.invalid[GST_AV1_REF_LAST_FRAME + i];
|
|
}
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_av1_dec_start_picture (GstAV1Decoder * decoder, GstAV1Picture * picture,
|
|
GstAV1Dpb * dpb)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstAV1FrameHeaderOBU *frame_header = &picture->frame_hdr;
|
|
GstAV1SequenceHeaderOBU *seq_header = &self->seq;
|
|
VADecPictureParameterBufferAV1 pic_param = { 0, };
|
|
GstVaDecodePicture *va_pic;
|
|
guint i;
|
|
|
|
va_pic = gst_av1_picture_get_user_data (picture);
|
|
g_assert (va_pic);
|
|
|
|
/* *INDENT-OFF* */
|
|
pic_param = (VADecPictureParameterBufferAV1){
|
|
.profile = seq_header->seq_profile,
|
|
.order_hint_bits_minus_1 = seq_header->order_hint_bits_minus_1,
|
|
.matrix_coefficients = seq_header->color_config.matrix_coefficients,
|
|
.seq_info_fields.fields = {
|
|
.still_picture = seq_header->still_picture,
|
|
.use_128x128_superblock = seq_header->use_128x128_superblock,
|
|
.enable_filter_intra = seq_header->enable_filter_intra,
|
|
.enable_intra_edge_filter = seq_header->enable_intra_edge_filter,
|
|
.enable_interintra_compound = seq_header->enable_interintra_compound,
|
|
.enable_masked_compound = seq_header->enable_masked_compound,
|
|
.enable_dual_filter = seq_header->enable_dual_filter,
|
|
.enable_order_hint = seq_header->enable_order_hint,
|
|
.enable_jnt_comp = seq_header->enable_jnt_comp,
|
|
.enable_cdef = seq_header->enable_cdef,
|
|
.mono_chrome = seq_header->color_config.mono_chrome,
|
|
.color_range = seq_header->color_config.color_range,
|
|
.subsampling_x = seq_header->color_config.subsampling_x,
|
|
.subsampling_y = seq_header->color_config.subsampling_y,
|
|
.film_grain_params_present = seq_header->film_grain_params_present,
|
|
},
|
|
.anchor_frames_num = 0,
|
|
.anchor_frames_list = NULL,
|
|
.frame_width_minus1 = frame_header->upscaled_width - 1,
|
|
.frame_height_minus1 = frame_header->frame_height - 1,
|
|
.output_frame_width_in_tiles_minus_1 = 0,
|
|
.output_frame_height_in_tiles_minus_1 = 0,
|
|
.order_hint = frame_header->order_hint,
|
|
/* Segmentation */
|
|
.seg_info.segment_info_fields.bits = {
|
|
.enabled = frame_header->segmentation_params.segmentation_enabled,
|
|
.update_map = frame_header->segmentation_params.segmentation_update_map,
|
|
.temporal_update =
|
|
frame_header->segmentation_params.segmentation_temporal_update,
|
|
.update_data =
|
|
frame_header->segmentation_params.segmentation_update_data,
|
|
},
|
|
/* FilmGrain */
|
|
.film_grain_info = {
|
|
.film_grain_info_fields.bits = {
|
|
.apply_grain = frame_header->film_grain_params.apply_grain,
|
|
.chroma_scaling_from_luma =
|
|
frame_header->film_grain_params.chroma_scaling_from_luma,
|
|
.grain_scaling_minus_8 =
|
|
frame_header->film_grain_params.grain_scaling_minus_8,
|
|
.ar_coeff_lag = frame_header->film_grain_params.ar_coeff_lag,
|
|
.ar_coeff_shift_minus_6 =
|
|
frame_header->film_grain_params.ar_coeff_shift_minus_6,
|
|
.grain_scale_shift = frame_header->film_grain_params.grain_scale_shift,
|
|
.overlap_flag = frame_header->film_grain_params.overlap_flag,
|
|
.clip_to_restricted_range =
|
|
frame_header->film_grain_params.clip_to_restricted_range,
|
|
},
|
|
.grain_seed = frame_header->film_grain_params.grain_seed,
|
|
.cb_mult = frame_header->film_grain_params.cb_mult,
|
|
.cb_luma_mult = frame_header->film_grain_params.cb_luma_mult,
|
|
.cb_offset = frame_header->film_grain_params.cb_offset,
|
|
.cr_mult = frame_header->film_grain_params.cr_mult,
|
|
.cr_luma_mult = frame_header->film_grain_params.cr_luma_mult,
|
|
.cr_offset = frame_header->film_grain_params.cr_offset,
|
|
},
|
|
.tile_cols = frame_header->tile_info.tile_cols,
|
|
.tile_rows = frame_header->tile_info.tile_rows,
|
|
.context_update_tile_id = frame_header->tile_info.context_update_tile_id,
|
|
.pic_info_fields.bits = {
|
|
.frame_type = frame_header->frame_type,
|
|
.show_frame = frame_header->show_frame,
|
|
.showable_frame = frame_header->showable_frame,
|
|
.error_resilient_mode = frame_header->error_resilient_mode,
|
|
.disable_cdf_update = frame_header->disable_cdf_update,
|
|
.allow_screen_content_tools = frame_header->allow_screen_content_tools,
|
|
.force_integer_mv = frame_header->force_integer_mv,
|
|
.allow_intrabc = frame_header->allow_intrabc,
|
|
.use_superres = frame_header->use_superres,
|
|
.allow_high_precision_mv = frame_header->allow_high_precision_mv,
|
|
.is_motion_mode_switchable = frame_header->is_motion_mode_switchable,
|
|
.use_ref_frame_mvs = frame_header->use_ref_frame_mvs,
|
|
.disable_frame_end_update_cdf =
|
|
frame_header->disable_frame_end_update_cdf,
|
|
.uniform_tile_spacing_flag =
|
|
frame_header->tile_info.uniform_tile_spacing_flag,
|
|
.allow_warped_motion = frame_header->allow_warped_motion,
|
|
},
|
|
.superres_scale_denominator = frame_header->superres_denom,
|
|
.interp_filter = frame_header->interpolation_filter,
|
|
/* loop filter */
|
|
.loop_filter_info_fields.bits = {
|
|
.sharpness_level =
|
|
frame_header->loop_filter_params.loop_filter_sharpness,
|
|
.mode_ref_delta_enabled =
|
|
frame_header->loop_filter_params.loop_filter_delta_enabled,
|
|
.mode_ref_delta_update =
|
|
frame_header->loop_filter_params.loop_filter_delta_update,
|
|
},
|
|
.mode_control_fields.bits = {
|
|
.delta_lf_present_flag =
|
|
frame_header->loop_filter_params.delta_lf_present,
|
|
.log2_delta_lf_res = frame_header->loop_filter_params.delta_lf_res,
|
|
.delta_lf_multi = frame_header->loop_filter_params.delta_lf_multi,
|
|
.delta_q_present_flag =
|
|
frame_header->quantization_params.delta_q_present,
|
|
.log2_delta_q_res = frame_header->quantization_params.delta_q_res,
|
|
.tx_mode = frame_header->tx_mode,
|
|
.reference_select = frame_header->reference_select,
|
|
.reduced_tx_set_used = frame_header->reduced_tx_set,
|
|
.skip_mode_present = frame_header->skip_mode_present,
|
|
},
|
|
/* quantization */
|
|
.base_qindex = frame_header->quantization_params.base_q_idx,
|
|
.y_dc_delta_q = frame_header->quantization_params.delta_q_y_dc,
|
|
.u_dc_delta_q = frame_header->quantization_params.delta_q_u_dc,
|
|
.u_ac_delta_q = frame_header->quantization_params.delta_q_u_ac,
|
|
.v_dc_delta_q = frame_header->quantization_params.delta_q_v_dc,
|
|
.v_ac_delta_q = frame_header->quantization_params.delta_q_v_ac,
|
|
/* loop restoration */
|
|
.loop_restoration_fields.bits = {
|
|
.yframe_restoration_type =
|
|
frame_header->loop_restoration_params.frame_restoration_type[0],
|
|
.cbframe_restoration_type =
|
|
frame_header->loop_restoration_params.frame_restoration_type[1],
|
|
.crframe_restoration_type =
|
|
frame_header->loop_restoration_params.frame_restoration_type[2],
|
|
.lr_unit_shift = frame_header->loop_restoration_params.lr_unit_shift,
|
|
.lr_uv_shift = frame_header->loop_restoration_params.lr_uv_shift,
|
|
},
|
|
};
|
|
/* *INDENT-ON* */
|
|
|
|
if (seq_header->bit_depth == 8) {
|
|
pic_param.bit_depth_idx = 0;
|
|
} else if (seq_header->bit_depth == 10) {
|
|
pic_param.bit_depth_idx = 1;
|
|
} else if (seq_header->bit_depth == 12) {
|
|
pic_param.bit_depth_idx = 2;
|
|
} else {
|
|
g_assert_not_reached ();
|
|
}
|
|
|
|
if (frame_header->film_grain_params.apply_grain) {
|
|
pic_param.current_frame = gst_va_decode_picture_get_aux_surface (va_pic);
|
|
pic_param.current_display_picture =
|
|
gst_va_decode_picture_get_surface (va_pic);
|
|
} else {
|
|
pic_param.current_frame = gst_va_decode_picture_get_surface (va_pic);
|
|
pic_param.current_display_picture = VA_INVALID_SURFACE;
|
|
}
|
|
|
|
for (i = 0; i < GST_AV1_NUM_REF_FRAMES; i++) {
|
|
if (dpb->pic_list[i]) {
|
|
if (dpb->pic_list[i]->apply_grain) {
|
|
pic_param.ref_frame_map[i] = gst_va_decode_picture_get_aux_surface
|
|
(gst_av1_picture_get_user_data (dpb->pic_list[i]));
|
|
} else {
|
|
pic_param.ref_frame_map[i] = gst_va_decode_picture_get_surface
|
|
(gst_av1_picture_get_user_data (dpb->pic_list[i]));
|
|
}
|
|
} else {
|
|
pic_param.ref_frame_map[i] = VA_INVALID_SURFACE;
|
|
}
|
|
}
|
|
for (i = 0; i < GST_AV1_REFS_PER_FRAME; i++) {
|
|
pic_param.ref_frame_idx[i] = frame_header->ref_frame_idx[i];
|
|
}
|
|
pic_param.primary_ref_frame = frame_header->primary_ref_frame;
|
|
|
|
_setup_segment_info (&pic_param, frame_header);
|
|
_setup_film_grain_info (&pic_param, frame_header);
|
|
|
|
for (i = 0; i < 63; i++) {
|
|
pic_param.width_in_sbs_minus_1[i] =
|
|
frame_header->tile_info.width_in_sbs_minus_1[i];
|
|
pic_param.height_in_sbs_minus_1[i] =
|
|
frame_header->tile_info.height_in_sbs_minus_1[i];
|
|
}
|
|
|
|
_setup_loop_filter_info (&pic_param, frame_header);
|
|
_setup_quantization_info (&pic_param, frame_header);
|
|
_setup_cdef_info (&pic_param, frame_header, seq_header->num_planes);
|
|
_setup_global_motion_info (&pic_param, frame_header);
|
|
|
|
if (!gst_va_decoder_add_param_buffer (base->decoder, va_pic,
|
|
VAPictureParameterBufferType, &pic_param, sizeof (pic_param)))
|
|
return GST_FLOW_ERROR;
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_av1_dec_decode_tile (GstAV1Decoder * decoder, GstAV1Picture * picture,
|
|
GstAV1Tile * tile)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstAV1TileGroupOBU *tile_group = &tile->tile_group;
|
|
GstVaDecodePicture *va_pic;
|
|
guint i;
|
|
VASliceParameterBufferAV1 slice_param[GST_AV1_MAX_TILE_COUNT] = { 0, };
|
|
|
|
GST_TRACE_OBJECT (self, "-");
|
|
|
|
for (i = 0; i < tile_group->tg_end - tile_group->tg_start + 1; i++) {
|
|
slice_param[i].slice_data_size =
|
|
tile_group->entry[tile_group->tg_start + i].tile_size;
|
|
slice_param[i].slice_data_offset =
|
|
tile_group->entry[tile_group->tg_start + i].tile_offset;
|
|
slice_param[i].tile_row =
|
|
tile_group->entry[tile_group->tg_start + i].tile_row;
|
|
slice_param[i].tile_column =
|
|
tile_group->entry[tile_group->tg_start + i].tile_col;
|
|
slice_param[i].slice_data_flag = 0;
|
|
}
|
|
|
|
va_pic = gst_av1_picture_get_user_data (picture);
|
|
|
|
if (!gst_va_decoder_add_slice_buffer_with_n_params (base->decoder, va_pic,
|
|
slice_param, sizeof (VASliceParameterBufferAV1), i, tile->obu.data,
|
|
tile->obu.obu_size)) {
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_av1_dec_end_picture (GstAV1Decoder * decoder, GstAV1Picture * picture)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVaDecodePicture *va_pic;
|
|
|
|
GST_LOG_OBJECT (self, "end picture %p, (system_frame_number %d)",
|
|
picture, GST_CODEC_PICTURE_FRAME_NUMBER (picture));
|
|
|
|
va_pic = gst_av1_picture_get_user_data (picture);
|
|
|
|
if (!gst_va_decoder_decode_with_aux_surface (base->decoder, va_pic,
|
|
picture->apply_grain)) {
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_va_av1_dec_output_picture (GstAV1Decoder * decoder,
|
|
GstVideoCodecFrame * frame, GstAV1Picture * picture)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
GstVaBaseDec *base = GST_VA_BASE_DEC (decoder);
|
|
GstVideoDecoder *vdec = GST_VIDEO_DECODER (decoder);
|
|
GstCodecPicture *codec_picture = GST_CODEC_PICTURE (picture);
|
|
gboolean ret;
|
|
|
|
g_assert (picture->frame_hdr.show_frame ||
|
|
picture->frame_hdr.show_existing_frame);
|
|
|
|
GST_LOG_OBJECT (self,
|
|
"Outputting picture %p (system_frame_number %d)",
|
|
picture, codec_picture->system_frame_number);
|
|
|
|
if (picture->frame_hdr.show_existing_frame) {
|
|
GstVaDecodePicture *pic;
|
|
|
|
g_assert (!frame->output_buffer);
|
|
pic = gst_av1_picture_get_user_data (picture);
|
|
frame->output_buffer = gst_buffer_ref (pic->gstbuffer);
|
|
}
|
|
|
|
ret = gst_va_base_dec_process_output (base,
|
|
frame, codec_picture->discont_state, 0);
|
|
gst_av1_picture_unref (picture);
|
|
|
|
if (ret)
|
|
return gst_video_decoder_finish_frame (vdec, frame);
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
static gboolean
|
|
gst_va_av1_dec_start (GstVideoDecoder * decoder)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
|
|
self->preferred_format = GST_VIDEO_FORMAT_UNKNOWN;
|
|
|
|
return GST_VIDEO_DECODER_CLASS (parent_class)->start (decoder);
|
|
}
|
|
|
|
static gboolean
|
|
gst_va_av1_dec_close (GstVideoDecoder * decoder)
|
|
{
|
|
GstVaAV1Dec *self = GST_VA_AV1_DEC (decoder);
|
|
|
|
_clear_internal_pool (self);
|
|
|
|
return gst_va_base_dec_close (GST_VIDEO_DECODER (decoder));
|
|
}
|
|
|
|
static void
|
|
gst_va_av1_dec_init (GTypeInstance * instance, gpointer g_class)
|
|
{
|
|
gst_va_base_dec_init (GST_VA_BASE_DEC (instance), GST_CAT_DEFAULT);
|
|
}
|
|
|
|
static void
|
|
gst_va_av1_dec_dispose (GObject * object)
|
|
{
|
|
gst_va_base_dec_close (GST_VIDEO_DECODER (object));
|
|
G_OBJECT_CLASS (parent_class)->dispose (object);
|
|
}
|
|
|
|
static void
|
|
gst_va_av1_dec_class_init (gpointer g_class, gpointer class_data)
|
|
{
|
|
GstCaps *src_doc_caps, *sink_doc_caps;
|
|
GObjectClass *gobject_class = G_OBJECT_CLASS (g_class);
|
|
GstElementClass *element_class = GST_ELEMENT_CLASS (g_class);
|
|
GstAV1DecoderClass *av1decoder_class = GST_AV1_DECODER_CLASS (g_class);
|
|
GstVideoDecoderClass *decoder_class = GST_VIDEO_DECODER_CLASS (g_class);
|
|
struct CData *cdata = class_data;
|
|
gchar *long_name;
|
|
|
|
if (cdata->description) {
|
|
long_name = g_strdup_printf ("VA-API AV1 Decoder in %s",
|
|
cdata->description);
|
|
} else {
|
|
long_name = g_strdup ("VA-API AV1 Decoder");
|
|
}
|
|
|
|
gst_element_class_set_metadata (element_class, long_name,
|
|
"Codec/Decoder/Video/Hardware",
|
|
"VA-API based AV1 video decoder", "He Junyan <junyan.he@intel.com>");
|
|
|
|
sink_doc_caps = gst_caps_from_string (sink_caps_str);
|
|
src_doc_caps = gst_caps_from_string (src_caps_str);
|
|
|
|
parent_class = g_type_class_peek_parent (g_class);
|
|
|
|
/**
|
|
* GstVaAV1Dec:device-path:
|
|
*
|
|
* It shows the DRM device path used for the VA operation, if any.
|
|
*
|
|
* Since: 1.22
|
|
*/
|
|
gst_va_base_dec_class_init (GST_VA_BASE_DEC_CLASS (g_class), AV1,
|
|
cdata->render_device_path, cdata->sink_caps, cdata->src_caps,
|
|
src_doc_caps, sink_doc_caps);
|
|
|
|
gobject_class->dispose = gst_va_av1_dec_dispose;
|
|
|
|
decoder_class->getcaps = GST_DEBUG_FUNCPTR (gst_va_av1_dec_getcaps);
|
|
decoder_class->negotiate = GST_DEBUG_FUNCPTR (gst_va_av1_dec_negotiate);
|
|
decoder_class->close = GST_DEBUG_FUNCPTR (gst_va_av1_dec_close);
|
|
decoder_class->start = GST_DEBUG_FUNCPTR (gst_va_av1_dec_start);
|
|
|
|
av1decoder_class->new_sequence =
|
|
GST_DEBUG_FUNCPTR (gst_va_av1_dec_new_sequence);
|
|
av1decoder_class->new_picture =
|
|
GST_DEBUG_FUNCPTR (gst_va_av1_dec_new_picture);
|
|
av1decoder_class->duplicate_picture =
|
|
GST_DEBUG_FUNCPTR (gst_va_av1_dec_duplicate_picture);
|
|
av1decoder_class->start_picture =
|
|
GST_DEBUG_FUNCPTR (gst_va_av1_dec_start_picture);
|
|
av1decoder_class->decode_tile =
|
|
GST_DEBUG_FUNCPTR (gst_va_av1_dec_decode_tile);
|
|
av1decoder_class->end_picture =
|
|
GST_DEBUG_FUNCPTR (gst_va_av1_dec_end_picture);
|
|
av1decoder_class->output_picture =
|
|
GST_DEBUG_FUNCPTR (gst_va_av1_dec_output_picture);
|
|
|
|
g_free (long_name);
|
|
g_free (cdata->description);
|
|
g_free (cdata->render_device_path);
|
|
gst_caps_unref (cdata->src_caps);
|
|
gst_caps_unref (cdata->sink_caps);
|
|
g_free (cdata);
|
|
}
|
|
|
|
static gpointer
|
|
_register_debug_category (gpointer data)
|
|
{
|
|
GST_DEBUG_CATEGORY_INIT (gst_va_av1dec_debug, "vaav1dec", 0,
|
|
"VA AV1 decoder");
|
|
|
|
return NULL;
|
|
}
|
|
|
|
gboolean
|
|
gst_va_av1_dec_register (GstPlugin * plugin, GstVaDevice * device,
|
|
GstCaps * sink_caps, GstCaps * src_caps, guint rank)
|
|
{
|
|
static GOnce debug_once = G_ONCE_INIT;
|
|
GType type;
|
|
GTypeInfo type_info = {
|
|
.class_size = sizeof (GstVaAV1DecClass),
|
|
.class_init = gst_va_av1_dec_class_init,
|
|
.instance_size = sizeof (GstVaAV1Dec),
|
|
.instance_init = gst_va_av1_dec_init,
|
|
};
|
|
struct CData *cdata;
|
|
gboolean ret;
|
|
gchar *type_name, *feature_name;
|
|
|
|
g_return_val_if_fail (GST_IS_PLUGIN (plugin), FALSE);
|
|
g_return_val_if_fail (GST_IS_VA_DEVICE (device), FALSE);
|
|
g_return_val_if_fail (GST_IS_CAPS (sink_caps), FALSE);
|
|
g_return_val_if_fail (GST_IS_CAPS (src_caps), FALSE);
|
|
|
|
cdata = g_new (struct CData, 1);
|
|
cdata->description = NULL;
|
|
cdata->render_device_path = g_strdup (device->render_device_path);
|
|
cdata->sink_caps = _complete_sink_caps (sink_caps);
|
|
cdata->src_caps = gst_caps_ref (src_caps);
|
|
|
|
/* class data will be leaked if the element never gets instantiated */
|
|
GST_MINI_OBJECT_FLAG_SET (cdata->sink_caps,
|
|
GST_MINI_OBJECT_FLAG_MAY_BE_LEAKED);
|
|
GST_MINI_OBJECT_FLAG_SET (src_caps, GST_MINI_OBJECT_FLAG_MAY_BE_LEAKED);
|
|
|
|
type_info.class_data = cdata;
|
|
|
|
gst_va_create_feature_name (device, "GstVaAV1Dec", "GstVa%sAV1Dec",
|
|
&type_name, "vaav1dec", "va%sav1dec", &feature_name,
|
|
&cdata->description, &rank);
|
|
|
|
g_once (&debug_once, _register_debug_category, NULL);
|
|
|
|
type = g_type_register_static (GST_TYPE_AV1_DECODER,
|
|
type_name, &type_info, 0);
|
|
|
|
ret = gst_element_register (plugin, feature_name, rank, type);
|
|
|
|
g_free (type_name);
|
|
g_free (feature_name);
|
|
|
|
return ret;
|
|
}
|