mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-12-26 02:00:33 +00:00
67eea92044
Drop obsolete GST_VAAPI_IS_xxx() helper macros since we are no longer deriving from GObject and so those were only checking for whether the argument was NULL or not. This is now irrelevant, and even confusing to some extent, because we no longer have type checking. Note: this incurs more type checking (review) but the libgstvaapi is rather small, so this is manageable.
1057 lines
31 KiB
C
1057 lines
31 KiB
C
/*
|
|
* gstvaapidecoder.c - VA decoder abstraction
|
|
*
|
|
* Copyright (C) 2010-2011 Splitted-Desktop Systems
|
|
* Copyright (C) 2011-2013 Intel Corporation
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public License
|
|
* as published by the Free Software Foundation; either version 2.1
|
|
* of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free
|
|
* Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
|
|
* Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
/**
|
|
* SECTION:gstvaapidecoder
|
|
* @short_description: VA decoder abstraction
|
|
*/
|
|
|
|
#include "sysdeps.h"
|
|
#include "gstvaapicompat.h"
|
|
#include "gstvaapidecoder.h"
|
|
#include "gstvaapidecoder_priv.h"
|
|
#include "gstvaapiparser_frame.h"
|
|
#include "gstvaapisurfaceproxy_priv.h"
|
|
#include "gstvaapiutils.h"
|
|
|
|
#define DEBUG 1
|
|
#include "gstvaapidebug.h"
|
|
|
|
static void
|
|
drop_frame(GstVaapiDecoder *decoder, GstVideoCodecFrame *frame);
|
|
|
|
static void
|
|
parser_state_finalize(GstVaapiParserState *ps)
|
|
{
|
|
if (ps->input_adapter) {
|
|
gst_adapter_clear(ps->input_adapter);
|
|
g_object_unref(ps->input_adapter);
|
|
ps->input_adapter = NULL;
|
|
}
|
|
|
|
if (ps->output_adapter) {
|
|
gst_adapter_clear(ps->output_adapter);
|
|
g_object_unref(ps->output_adapter);
|
|
ps->output_adapter = NULL;
|
|
}
|
|
|
|
if (ps->next_unit_pending) {
|
|
gst_vaapi_decoder_unit_clear(&ps->next_unit);
|
|
ps->next_unit_pending = FALSE;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
parser_state_init(GstVaapiParserState *ps)
|
|
{
|
|
memset(ps, 0, sizeof(*ps));
|
|
|
|
ps->input_adapter = gst_adapter_new();
|
|
if (!ps->input_adapter)
|
|
return FALSE;
|
|
|
|
ps->output_adapter = gst_adapter_new();
|
|
if (!ps->output_adapter)
|
|
return FALSE;
|
|
return TRUE;
|
|
}
|
|
|
|
static void
|
|
parser_state_prepare(GstVaapiParserState *ps, GstAdapter *adapter)
|
|
{
|
|
/* XXX: check we really have a continuity from the previous call */
|
|
if (ps->current_adapter != adapter)
|
|
goto reset;
|
|
return;
|
|
|
|
reset:
|
|
ps->current_adapter = adapter;
|
|
ps->input_offset2 = -1;
|
|
}
|
|
|
|
static gboolean
|
|
push_buffer(GstVaapiDecoder *decoder, GstBuffer *buffer)
|
|
{
|
|
if (!buffer) {
|
|
buffer = gst_buffer_new();
|
|
if (!buffer)
|
|
return FALSE;
|
|
GST_BUFFER_FLAG_SET(buffer, GST_BUFFER_FLAG_EOS);
|
|
}
|
|
|
|
GST_DEBUG("queue encoded data buffer %p (%d bytes)",
|
|
buffer, gst_buffer_get_size(buffer));
|
|
|
|
g_async_queue_push(decoder->buffers, buffer);
|
|
return TRUE;
|
|
}
|
|
|
|
static GstBuffer *
|
|
pop_buffer(GstVaapiDecoder *decoder)
|
|
{
|
|
GstBuffer *buffer;
|
|
|
|
buffer = g_async_queue_try_pop(decoder->buffers);
|
|
if (!buffer)
|
|
return NULL;
|
|
|
|
GST_DEBUG("dequeue buffer %p for decoding (%d bytes)",
|
|
buffer, gst_buffer_get_size(buffer));
|
|
|
|
return buffer;
|
|
}
|
|
|
|
static GstVaapiDecoderStatus
|
|
do_parse(GstVaapiDecoder *decoder,
|
|
GstVideoCodecFrame *base_frame, GstAdapter *adapter, gboolean at_eos,
|
|
guint *got_unit_size_ptr, gboolean *got_frame_ptr)
|
|
{
|
|
GstVaapiParserState * const ps = &decoder->parser_state;
|
|
GstVaapiParserFrame *frame;
|
|
GstVaapiDecoderUnit *unit;
|
|
GstVaapiDecoderStatus status;
|
|
|
|
*got_unit_size_ptr = 0;
|
|
*got_frame_ptr = FALSE;
|
|
|
|
frame = gst_video_codec_frame_get_user_data(base_frame);
|
|
if (!frame) {
|
|
GstVideoCodecState * const codec_state = decoder->codec_state;
|
|
frame = gst_vaapi_parser_frame_new(codec_state->info.width,
|
|
codec_state->info.height);
|
|
if (!frame)
|
|
return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED;
|
|
gst_video_codec_frame_set_user_data(base_frame,
|
|
frame, (GDestroyNotify)gst_vaapi_mini_object_unref);
|
|
}
|
|
|
|
parser_state_prepare(ps, adapter);
|
|
|
|
unit = &ps->next_unit;
|
|
if (ps->next_unit_pending) {
|
|
ps->next_unit_pending = FALSE;
|
|
goto got_unit;
|
|
}
|
|
gst_vaapi_decoder_unit_init(unit);
|
|
|
|
ps->current_frame = base_frame;
|
|
status = GST_VAAPI_DECODER_GET_CLASS(decoder)->parse(decoder,
|
|
adapter, at_eos, unit);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) {
|
|
if (at_eos && frame->units->len > 0 &&
|
|
status == GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA) {
|
|
/* XXX: assume the frame is complete at <EOS> */
|
|
*got_frame_ptr = TRUE;
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
return status;
|
|
}
|
|
|
|
if (GST_VAAPI_DECODER_UNIT_IS_FRAME_START(unit) && frame->units->len > 0) {
|
|
ps->next_unit_pending = TRUE;
|
|
*got_frame_ptr = TRUE;
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
|
|
got_unit:
|
|
gst_vaapi_parser_frame_append_unit(frame, unit);
|
|
*got_unit_size_ptr = unit->size;
|
|
*got_frame_ptr = GST_VAAPI_DECODER_UNIT_IS_FRAME_END(unit);
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
|
|
static GstVaapiDecoderStatus
|
|
do_decode_units(GstVaapiDecoder *decoder, GArray *units)
|
|
{
|
|
GstVaapiDecoderClass * const klass = GST_VAAPI_DECODER_GET_CLASS(decoder);
|
|
GstVaapiDecoderStatus status;
|
|
guint i;
|
|
|
|
for (i = 0; i < units->len; i++) {
|
|
GstVaapiDecoderUnit * const unit =
|
|
&g_array_index(units, GstVaapiDecoderUnit, i);
|
|
if (GST_VAAPI_DECODER_UNIT_IS_SKIPPED(unit))
|
|
continue;
|
|
status = klass->decode(decoder, unit);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS)
|
|
return status;
|
|
}
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
|
|
static GstVaapiDecoderStatus
|
|
do_decode_1(GstVaapiDecoder *decoder, GstVaapiParserFrame *frame)
|
|
{
|
|
GstVaapiDecoderClass * const klass = GST_VAAPI_DECODER_GET_CLASS(decoder);
|
|
GstVaapiDecoderStatus status;
|
|
|
|
if (frame->pre_units->len > 0) {
|
|
status = do_decode_units(decoder, frame->pre_units);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS)
|
|
return status;
|
|
}
|
|
|
|
if (frame->units->len > 0) {
|
|
if (klass->start_frame) {
|
|
GstVaapiDecoderUnit * const unit =
|
|
&g_array_index(frame->units, GstVaapiDecoderUnit, 0);
|
|
status = klass->start_frame(decoder, unit);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS)
|
|
return status;
|
|
}
|
|
|
|
status = do_decode_units(decoder, frame->units);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS)
|
|
return status;
|
|
|
|
if (klass->end_frame) {
|
|
status = klass->end_frame(decoder);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS)
|
|
return status;
|
|
}
|
|
}
|
|
|
|
if (frame->post_units->len > 0) {
|
|
status = do_decode_units(decoder, frame->post_units);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS)
|
|
return status;
|
|
}
|
|
|
|
/* Drop frame if there is no slice data unit in there */
|
|
if (G_UNLIKELY(frame->units->len == 0))
|
|
return GST_VAAPI_DECODER_STATUS_DROP_FRAME;
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
|
|
static inline GstVaapiDecoderStatus
|
|
do_decode(GstVaapiDecoder *decoder, GstVideoCodecFrame *base_frame)
|
|
{
|
|
GstVaapiParserState * const ps = &decoder->parser_state;
|
|
GstVaapiParserFrame * const frame = base_frame->user_data;
|
|
GstVaapiDecoderStatus status;
|
|
|
|
ps->current_frame = base_frame;
|
|
|
|
gst_vaapi_parser_frame_ref(frame);
|
|
status = do_decode_1(decoder, frame);
|
|
gst_vaapi_parser_frame_unref(frame);
|
|
|
|
switch ((guint)status) {
|
|
case GST_VAAPI_DECODER_STATUS_DROP_FRAME:
|
|
drop_frame(decoder, base_frame);
|
|
status = GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
break;
|
|
}
|
|
return status;
|
|
}
|
|
|
|
static inline GstVaapiDecoderStatus
|
|
do_flush(GstVaapiDecoder *decoder)
|
|
{
|
|
GstVaapiDecoderClass * const klass = GST_VAAPI_DECODER_GET_CLASS(decoder);
|
|
|
|
if (klass->flush)
|
|
return klass->flush(decoder);
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
|
|
static GstVaapiDecoderStatus
|
|
decode_step(GstVaapiDecoder *decoder)
|
|
{
|
|
GstVaapiParserState * const ps = &decoder->parser_state;
|
|
GstVaapiDecoderStatus status;
|
|
GstBuffer *buffer;
|
|
gboolean got_frame;
|
|
guint got_unit_size, input_size;
|
|
|
|
status = gst_vaapi_decoder_check_status(decoder);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS)
|
|
return status;
|
|
|
|
/* Fill adapter with all buffers we have in the queue */
|
|
for (;;) {
|
|
buffer = pop_buffer(decoder);
|
|
if (!buffer)
|
|
break;
|
|
|
|
ps->at_eos = GST_BUFFER_IS_EOS(buffer);
|
|
if (!ps->at_eos)
|
|
gst_adapter_push(ps->input_adapter, buffer);
|
|
}
|
|
|
|
/* Parse and decode all decode units */
|
|
input_size = gst_adapter_available(ps->input_adapter);
|
|
if (input_size == 0) {
|
|
if (ps->at_eos)
|
|
return GST_VAAPI_DECODER_STATUS_END_OF_STREAM;
|
|
return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA;
|
|
}
|
|
|
|
do {
|
|
if (!ps->current_frame) {
|
|
ps->current_frame = g_slice_new0(GstVideoCodecFrame);
|
|
if (!ps->current_frame)
|
|
return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED;
|
|
ps->current_frame->ref_count = 1;
|
|
}
|
|
|
|
status = do_parse(decoder, ps->current_frame, ps->input_adapter,
|
|
ps->at_eos, &got_unit_size, &got_frame);
|
|
GST_DEBUG("parse frame (status = %d)", status);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS) {
|
|
if (status == GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA && ps->at_eos)
|
|
status = GST_VAAPI_DECODER_STATUS_END_OF_STREAM;
|
|
break;
|
|
}
|
|
|
|
if (got_unit_size > 0) {
|
|
buffer = gst_adapter_take_buffer(ps->input_adapter, got_unit_size);
|
|
input_size -= got_unit_size;
|
|
|
|
if (gst_adapter_available(ps->output_adapter) == 0) {
|
|
ps->current_frame->pts =
|
|
gst_adapter_prev_timestamp(ps->input_adapter, NULL);
|
|
}
|
|
gst_adapter_push(ps->output_adapter, buffer);
|
|
}
|
|
|
|
if (got_frame) {
|
|
ps->current_frame->input_buffer = gst_adapter_take_buffer(
|
|
ps->output_adapter,
|
|
gst_adapter_available(ps->output_adapter));
|
|
|
|
status = do_decode(decoder, ps->current_frame);
|
|
GST_DEBUG("decode frame (status = %d)", status);
|
|
|
|
gst_video_codec_frame_unref(ps->current_frame);
|
|
ps->current_frame = NULL;
|
|
break;
|
|
}
|
|
} while (input_size > 0);
|
|
return status;
|
|
}
|
|
|
|
static void
|
|
drop_frame(GstVaapiDecoder *decoder, GstVideoCodecFrame *frame)
|
|
{
|
|
GST_DEBUG("drop frame %d", frame->system_frame_number);
|
|
|
|
/* no surface proxy */
|
|
gst_video_codec_frame_set_user_data(frame, NULL, NULL);
|
|
|
|
frame->pts = GST_CLOCK_TIME_NONE;
|
|
GST_VIDEO_CODEC_FRAME_FLAG_SET(frame,
|
|
GST_VIDEO_CODEC_FRAME_FLAG_DECODE_ONLY);
|
|
|
|
g_async_queue_push(decoder->frames, gst_video_codec_frame_ref(frame));
|
|
}
|
|
|
|
static inline void
|
|
push_frame(GstVaapiDecoder *decoder, GstVideoCodecFrame *frame)
|
|
{
|
|
GstVaapiSurfaceProxy * const proxy = frame->user_data;
|
|
|
|
GST_DEBUG("queue decoded surface %" GST_VAAPI_ID_FORMAT,
|
|
GST_VAAPI_ID_ARGS(GST_VAAPI_SURFACE_PROXY_SURFACE_ID(proxy)));
|
|
|
|
g_async_queue_push(decoder->frames, gst_video_codec_frame_ref(frame));
|
|
}
|
|
|
|
static inline GstVideoCodecFrame *
|
|
pop_frame(GstVaapiDecoder *decoder, guint64 timeout)
|
|
{
|
|
GstVideoCodecFrame *frame;
|
|
GstVaapiSurfaceProxy *proxy;
|
|
|
|
if (G_LIKELY(timeout > 0))
|
|
frame = g_async_queue_timeout_pop(decoder->frames, timeout);
|
|
else
|
|
frame = g_async_queue_try_pop(decoder->frames);
|
|
if (!frame)
|
|
return NULL;
|
|
|
|
proxy = frame->user_data;
|
|
GST_DEBUG("dequeue decoded surface %" GST_VAAPI_ID_FORMAT,
|
|
GST_VAAPI_ID_ARGS(GST_VAAPI_SURFACE_PROXY_SURFACE_ID(proxy)));
|
|
|
|
return frame;
|
|
}
|
|
|
|
static gboolean
|
|
set_caps(GstVaapiDecoder *decoder, const GstCaps *caps)
|
|
{
|
|
GstVideoCodecState * const codec_state = decoder->codec_state;
|
|
GstStructure * const structure = gst_caps_get_structure(caps, 0);
|
|
GstVaapiProfile profile;
|
|
const GValue *v_codec_data;
|
|
|
|
profile = gst_vaapi_profile_from_caps(caps);
|
|
if (!profile)
|
|
return FALSE;
|
|
|
|
decoder->codec = gst_vaapi_profile_get_codec(profile);
|
|
if (!decoder->codec)
|
|
return FALSE;
|
|
|
|
if (!gst_video_info_from_caps(&codec_state->info, caps))
|
|
return FALSE;
|
|
|
|
codec_state->caps = gst_caps_copy(caps);
|
|
|
|
v_codec_data = gst_structure_get_value(structure, "codec_data");
|
|
if (v_codec_data)
|
|
gst_buffer_replace(&codec_state->codec_data,
|
|
gst_value_get_buffer(v_codec_data));
|
|
return TRUE;
|
|
}
|
|
|
|
static inline GstCaps *
|
|
get_caps(GstVaapiDecoder *decoder)
|
|
{
|
|
return GST_VAAPI_DECODER_CODEC_STATE(decoder)->caps;
|
|
}
|
|
|
|
static void
|
|
notify_codec_state_changed(GstVaapiDecoder *decoder)
|
|
{
|
|
if (decoder->codec_state_changed_func)
|
|
decoder->codec_state_changed_func(decoder, decoder->codec_state,
|
|
decoder->codec_state_changed_data);
|
|
}
|
|
|
|
void
|
|
gst_vaapi_decoder_finalize(GstVaapiDecoder *decoder)
|
|
{
|
|
const GstVaapiDecoderClass * const klass =
|
|
GST_VAAPI_DECODER_GET_CLASS(decoder);
|
|
|
|
if (klass->destroy)
|
|
klass->destroy(decoder);
|
|
|
|
gst_video_codec_state_unref(decoder->codec_state);
|
|
decoder->codec_state = NULL;
|
|
|
|
parser_state_finalize(&decoder->parser_state);
|
|
|
|
if (decoder->buffers) {
|
|
g_async_queue_unref(decoder->buffers);
|
|
decoder->buffers = NULL;
|
|
}
|
|
|
|
if (decoder->frames) {
|
|
g_async_queue_unref(decoder->frames);
|
|
decoder->frames = NULL;
|
|
}
|
|
|
|
gst_vaapi_object_replace(&decoder->context, NULL);
|
|
decoder->va_context = VA_INVALID_ID;
|
|
|
|
gst_vaapi_display_replace(&decoder->display, NULL);
|
|
decoder->va_display = NULL;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vaapi_decoder_init(GstVaapiDecoder *decoder, GstVaapiDisplay *display,
|
|
GstCaps *caps)
|
|
{
|
|
const GstVaapiDecoderClass * const klass =
|
|
GST_VAAPI_DECODER_GET_CLASS(decoder);
|
|
GstVideoCodecState *codec_state;
|
|
guint sub_size;
|
|
|
|
parser_state_init(&decoder->parser_state);
|
|
|
|
codec_state = g_slice_new0(GstVideoCodecState);
|
|
codec_state->ref_count = 1;
|
|
gst_video_info_init(&codec_state->info);
|
|
|
|
decoder->user_data = NULL;
|
|
decoder->display = gst_vaapi_display_ref(display);
|
|
decoder->va_display = GST_VAAPI_DISPLAY_VADISPLAY(display);
|
|
decoder->context = NULL;
|
|
decoder->va_context = VA_INVALID_ID;
|
|
decoder->codec = 0;
|
|
decoder->codec_state = codec_state;
|
|
decoder->codec_state_changed_func = NULL;
|
|
decoder->codec_state_changed_data = NULL;
|
|
|
|
decoder->buffers = g_async_queue_new_full((GDestroyNotify)gst_buffer_unref);
|
|
decoder->frames = g_async_queue_new_full((GDestroyNotify)
|
|
gst_video_codec_frame_unref);
|
|
|
|
if (!set_caps(decoder, caps))
|
|
return FALSE;
|
|
|
|
sub_size = GST_VAAPI_MINI_OBJECT_CLASS(klass)->size - sizeof(*decoder);
|
|
if (sub_size > 0)
|
|
memset(((guchar *)decoder) + sizeof(*decoder), 0, sub_size);
|
|
|
|
if (klass->create && !klass->create(decoder))
|
|
return FALSE;
|
|
return TRUE;
|
|
}
|
|
|
|
GstVaapiDecoder *
|
|
gst_vaapi_decoder_new(const GstVaapiDecoderClass *klass,
|
|
GstVaapiDisplay *display, GstCaps *caps)
|
|
{
|
|
GstVaapiDecoder *decoder;
|
|
|
|
g_return_val_if_fail(display != NULL, NULL);
|
|
g_return_val_if_fail(GST_IS_CAPS(caps), NULL);
|
|
|
|
decoder = (GstVaapiDecoder *)
|
|
gst_vaapi_mini_object_new(GST_VAAPI_MINI_OBJECT_CLASS(klass));
|
|
if (!decoder)
|
|
return NULL;
|
|
|
|
if (!gst_vaapi_decoder_init(decoder, display, caps))
|
|
goto error;
|
|
return decoder;
|
|
|
|
error:
|
|
gst_vaapi_decoder_unref(decoder);
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_ref:
|
|
* @decoder: a #GstVaapiDecoder
|
|
*
|
|
* Atomically increases the reference count of the given @decoder by one.
|
|
*
|
|
* Returns: The same @decoder argument
|
|
*/
|
|
GstVaapiDecoder *
|
|
gst_vaapi_decoder_ref(GstVaapiDecoder *decoder)
|
|
{
|
|
return gst_vaapi_object_ref(decoder);
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_unref:
|
|
* @decoder: a #GstVaapiDecoder
|
|
*
|
|
* Atomically decreases the reference count of the @decoder by one. If
|
|
* the reference count reaches zero, the decoder will be free'd.
|
|
*/
|
|
void
|
|
gst_vaapi_decoder_unref(GstVaapiDecoder *decoder)
|
|
{
|
|
gst_vaapi_object_unref(decoder);
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_replace:
|
|
* @old_decoder_ptr: a pointer to a #GstVaapiDecoder
|
|
* @new_decoder: a #GstVaapiDecoder
|
|
*
|
|
* Atomically replaces the decoder decoder held in @old_decoder_ptr
|
|
* with @new_decoder. This means that @old_decoder_ptr shall reference
|
|
* a valid decoder. However, @new_decoder can be NULL.
|
|
*/
|
|
void
|
|
gst_vaapi_decoder_replace(GstVaapiDecoder **old_decoder_ptr,
|
|
GstVaapiDecoder *new_decoder)
|
|
{
|
|
gst_vaapi_object_replace(old_decoder_ptr, new_decoder);
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_get_user_data:
|
|
* @decoder: a #GstVaapiDecoder
|
|
*
|
|
* Retrieves the user-defined data associated with the @decoder, if any.
|
|
*
|
|
* Return value: the user-defined data associated with the @decoder
|
|
*/
|
|
gpointer
|
|
gst_vaapi_decoder_get_user_data(GstVaapiDecoder *decoder)
|
|
{
|
|
g_return_val_if_fail(decoder != NULL, NULL);
|
|
|
|
return decoder->user_data;
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_set_user_data:
|
|
* @decoder: a #GstVaapiDecoder
|
|
* @user_data: the pointer to user-defined data
|
|
*
|
|
* Associates user-defined @user_data to the @decoder. Retrieve the
|
|
* attached value with gst_vaapi_decoder_get_user_data() function.
|
|
*/
|
|
void
|
|
gst_vaapi_decoder_set_user_data(GstVaapiDecoder *decoder, gpointer user_data)
|
|
{
|
|
g_return_if_fail(decoder != NULL);
|
|
|
|
decoder->user_data = user_data;
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_get_codec:
|
|
* @decoder: a #GstVaapiDecoder
|
|
*
|
|
* Retrieves the @decoder codec type.
|
|
*
|
|
* Return value: the #GstVaapiCodec type for @decoder
|
|
*/
|
|
GstVaapiCodec
|
|
gst_vaapi_decoder_get_codec(GstVaapiDecoder *decoder)
|
|
{
|
|
g_return_val_if_fail(decoder != NULL, (GstVaapiCodec)0);
|
|
|
|
return decoder->codec;
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_get_codec_state:
|
|
* @decoder: a #GstVaapiDecoder
|
|
*
|
|
* Retrieves the @decoder codec state. The decoder owns the returned
|
|
* #GstVideoCodecState structure, so use gst_video_codec_state_ref()
|
|
* whenever necessary.
|
|
*
|
|
* Return value: the #GstVideoCodecState object for @decoder
|
|
*/
|
|
GstVideoCodecState *
|
|
gst_vaapi_decoder_get_codec_state(GstVaapiDecoder *decoder)
|
|
{
|
|
g_return_val_if_fail(decoder != NULL, NULL);
|
|
|
|
return GST_VAAPI_DECODER_CODEC_STATE(decoder);
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_set_codec_state_changed_func:
|
|
* @decoder: a #GstVaapiDecoder
|
|
* @func: the function to call when codec state changed
|
|
* @user_data: a pointer to user-defined data
|
|
*
|
|
* Sets @func as the function to call whenever the @decoder codec
|
|
* state changes.
|
|
*/
|
|
void
|
|
gst_vaapi_decoder_set_codec_state_changed_func(GstVaapiDecoder *decoder,
|
|
GstVaapiDecoderStateChangedFunc func, gpointer user_data)
|
|
{
|
|
g_return_if_fail(decoder != NULL);
|
|
|
|
decoder->codec_state_changed_func = func;
|
|
decoder->codec_state_changed_data = user_data;
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_get_caps:
|
|
* @decoder: a #GstVaapiDecoder
|
|
*
|
|
* Retrieves the @decoder caps. The decoder owns the returned caps, so
|
|
* use gst_caps_ref() whenever necessary.
|
|
*
|
|
* Return value: the @decoder caps
|
|
*/
|
|
GstCaps *
|
|
gst_vaapi_decoder_get_caps(GstVaapiDecoder *decoder)
|
|
{
|
|
return get_caps(decoder);
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_put_buffer:
|
|
* @decoder: a #GstVaapiDecoder
|
|
* @buf: a #GstBuffer
|
|
*
|
|
* Queues a #GstBuffer to the HW decoder. The decoder holds a
|
|
* reference to @buf.
|
|
*
|
|
* Caller can notify an End-Of-Stream with @buf set to %NULL. However,
|
|
* if an empty buffer is passed, i.e. a buffer with %NULL data pointer
|
|
* or size equals to zero, then the function ignores this buffer and
|
|
* returns %TRUE.
|
|
*
|
|
* Return value: %TRUE on success
|
|
*/
|
|
gboolean
|
|
gst_vaapi_decoder_put_buffer(GstVaapiDecoder *decoder, GstBuffer *buf)
|
|
{
|
|
g_return_val_if_fail(decoder != NULL, FALSE);
|
|
|
|
if (buf) {
|
|
if (gst_buffer_get_size(buf) == 0)
|
|
return TRUE;
|
|
buf = gst_buffer_ref(buf);
|
|
}
|
|
return push_buffer(decoder, buf);
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_get_surface:
|
|
* @decoder: a #GstVaapiDecoder
|
|
* @out_proxy_ptr: the next decoded surface as a #GstVaapiSurfaceProxy
|
|
*
|
|
* Flushes encoded buffers to the decoder and returns a decoded
|
|
* surface, if any.
|
|
*
|
|
* On successful return, *@out_proxy_ptr contains the decoded surface
|
|
* as a #GstVaapiSurfaceProxy. The caller owns this object, so
|
|
* gst_vaapi_surface_proxy_unref() shall be called after usage.
|
|
*
|
|
* Return value: a #GstVaapiDecoderStatus
|
|
*/
|
|
GstVaapiDecoderStatus
|
|
gst_vaapi_decoder_get_surface(GstVaapiDecoder *decoder,
|
|
GstVaapiSurfaceProxy **out_proxy_ptr)
|
|
{
|
|
GstVideoCodecFrame *frame;
|
|
GstVaapiDecoderStatus status;
|
|
|
|
g_return_val_if_fail(decoder != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
g_return_val_if_fail(out_proxy_ptr != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
|
|
do {
|
|
frame = pop_frame(decoder, 0);
|
|
while (frame) {
|
|
if (!GST_VIDEO_CODEC_FRAME_IS_DECODE_ONLY(frame)) {
|
|
GstVaapiSurfaceProxy * const proxy = frame->user_data;
|
|
proxy->timestamp = frame->pts;
|
|
proxy->duration = frame->duration;
|
|
*out_proxy_ptr = gst_vaapi_surface_proxy_ref(proxy);
|
|
gst_video_codec_frame_unref(frame);
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
gst_video_codec_frame_unref(frame);
|
|
frame = pop_frame(decoder, 0);
|
|
}
|
|
status = decode_step(decoder);
|
|
} while (status == GST_VAAPI_DECODER_STATUS_SUCCESS);
|
|
|
|
*out_proxy_ptr = NULL;
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_get_frame:
|
|
* @decoder: a #GstVaapiDecoder
|
|
* @out_frame_ptr: the next decoded frame as a #GstVideoCodecFrame
|
|
*
|
|
* On successful return, *@out_frame_ptr contains the next decoded
|
|
* frame available as a #GstVideoCodecFrame. The caller owns this
|
|
* object, so gst_video_codec_frame_unref() shall be called after
|
|
* usage. Otherwise, @GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA is
|
|
* returned if no decoded frame is available.
|
|
*
|
|
* The actual surface is available as a #GstVaapiSurfaceProxy attached
|
|
* to the user-data anchor of the output frame. Ownership of the proxy
|
|
* is transferred to the frame.
|
|
*
|
|
* This is equivalent to gst_vaapi_decoder_get_frame_with_timeout()
|
|
* with a timeout value of zero.
|
|
*
|
|
* Return value: a #GstVaapiDecoderStatus
|
|
*/
|
|
GstVaapiDecoderStatus
|
|
gst_vaapi_decoder_get_frame(GstVaapiDecoder *decoder,
|
|
GstVideoCodecFrame **out_frame_ptr)
|
|
{
|
|
return gst_vaapi_decoder_get_frame_with_timeout(decoder, out_frame_ptr, 0);
|
|
}
|
|
|
|
/**
|
|
* gst_vaapi_decoder_get_frame_with_timeout:
|
|
* @decoder: a #GstVaapiDecoder
|
|
* @out_frame_ptr: the next decoded frame as a #GstVideoCodecFrame
|
|
* @timeout: the number of microseconds to wait for the frame, at most
|
|
*
|
|
* On successful return, *@out_frame_ptr contains the next decoded
|
|
* frame available as a #GstVideoCodecFrame. The caller owns this
|
|
* object, so gst_video_codec_frame_unref() shall be called after
|
|
* usage. Otherwise, @GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA is
|
|
* returned if no decoded frame is available.
|
|
*
|
|
* The actual surface is available as a #GstVaapiSurfaceProxy attached
|
|
* to the user-data anchor of the output frame. Ownership of the proxy
|
|
* is transferred to the frame.
|
|
*
|
|
* Return value: a #GstVaapiDecoderStatus
|
|
*/
|
|
GstVaapiDecoderStatus
|
|
gst_vaapi_decoder_get_frame_with_timeout(GstVaapiDecoder *decoder,
|
|
GstVideoCodecFrame **out_frame_ptr, guint64 timeout)
|
|
{
|
|
GstVideoCodecFrame *out_frame;
|
|
|
|
g_return_val_if_fail(decoder != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
g_return_val_if_fail(out_frame_ptr != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
|
|
out_frame = pop_frame(decoder, timeout);
|
|
if (!out_frame)
|
|
return GST_VAAPI_DECODER_STATUS_ERROR_NO_DATA;
|
|
|
|
#if !GST_CHECK_VERSION(1,0,0)
|
|
if (!GST_VIDEO_CODEC_FRAME_IS_DECODE_ONLY(out_frame)) {
|
|
const guint flags = GST_VAAPI_SURFACE_PROXY_FLAGS(out_frame->user_data);
|
|
guint out_flags = 0;
|
|
|
|
if (flags & GST_VAAPI_SURFACE_PROXY_FLAG_TFF)
|
|
out_flags |= GST_VIDEO_CODEC_FRAME_FLAG_TFF;
|
|
if (flags & GST_VAAPI_SURFACE_PROXY_FLAG_RFF)
|
|
out_flags |= GST_VIDEO_CODEC_FRAME_FLAG_RFF;
|
|
if (flags & GST_VAAPI_SURFACE_PROXY_FLAG_ONEFIELD)
|
|
out_flags |= GST_VIDEO_CODEC_FRAME_FLAG_ONEFIELD;
|
|
GST_VIDEO_CODEC_FRAME_FLAG_SET(out_frame, out_flags);
|
|
}
|
|
#endif
|
|
|
|
*out_frame_ptr = out_frame;
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
|
|
void
|
|
gst_vaapi_decoder_set_picture_size(
|
|
GstVaapiDecoder *decoder,
|
|
guint width,
|
|
guint height
|
|
)
|
|
{
|
|
GstVideoCodecState * const codec_state = decoder->codec_state;
|
|
gboolean size_changed = FALSE;
|
|
|
|
if (codec_state->info.width != width) {
|
|
GST_DEBUG("picture width changed to %d", width);
|
|
codec_state->info.width = width;
|
|
gst_caps_set_simple(codec_state->caps,
|
|
"width", G_TYPE_INT, width, NULL);
|
|
size_changed = TRUE;
|
|
}
|
|
|
|
if (codec_state->info.height != height) {
|
|
GST_DEBUG("picture height changed to %d", height);
|
|
codec_state->info.height = height;
|
|
gst_caps_set_simple(codec_state->caps,
|
|
"height", G_TYPE_INT, height, NULL);
|
|
size_changed = TRUE;
|
|
}
|
|
|
|
if (size_changed)
|
|
notify_codec_state_changed(decoder);
|
|
}
|
|
|
|
void
|
|
gst_vaapi_decoder_set_framerate(
|
|
GstVaapiDecoder *decoder,
|
|
guint fps_n,
|
|
guint fps_d
|
|
)
|
|
{
|
|
GstVideoCodecState * const codec_state = decoder->codec_state;
|
|
|
|
if (!fps_n || !fps_d)
|
|
return;
|
|
|
|
if (codec_state->info.fps_n != fps_n || codec_state->info.fps_d != fps_d) {
|
|
GST_DEBUG("framerate changed to %u/%u", fps_n, fps_d);
|
|
codec_state->info.fps_n = fps_n;
|
|
codec_state->info.fps_d = fps_d;
|
|
gst_caps_set_simple(codec_state->caps,
|
|
"framerate", GST_TYPE_FRACTION, fps_n, fps_d, NULL);
|
|
notify_codec_state_changed(decoder);
|
|
}
|
|
}
|
|
|
|
void
|
|
gst_vaapi_decoder_set_pixel_aspect_ratio(
|
|
GstVaapiDecoder *decoder,
|
|
guint par_n,
|
|
guint par_d
|
|
)
|
|
{
|
|
GstVideoCodecState * const codec_state = decoder->codec_state;
|
|
|
|
if (!par_n || !par_d)
|
|
return;
|
|
|
|
if (codec_state->info.par_n != par_n || codec_state->info.par_d != par_d) {
|
|
GST_DEBUG("pixel-aspect-ratio changed to %u/%u", par_n, par_d);
|
|
codec_state->info.par_n = par_n;
|
|
codec_state->info.par_d = par_d;
|
|
gst_caps_set_simple(codec_state->caps,
|
|
"pixel-aspect-ratio", GST_TYPE_FRACTION, par_n, par_d, NULL);
|
|
notify_codec_state_changed(decoder);
|
|
}
|
|
}
|
|
|
|
static const gchar *
|
|
gst_interlace_mode_to_string(GstVideoInterlaceMode mode)
|
|
{
|
|
switch (mode) {
|
|
case GST_VIDEO_INTERLACE_MODE_PROGRESSIVE: return "progressive";
|
|
case GST_VIDEO_INTERLACE_MODE_INTERLEAVED: return "interleaved";
|
|
case GST_VIDEO_INTERLACE_MODE_MIXED: return "mixed";
|
|
}
|
|
return "<unknown>";
|
|
}
|
|
|
|
void
|
|
gst_vaapi_decoder_set_interlace_mode(GstVaapiDecoder *decoder,
|
|
GstVideoInterlaceMode mode)
|
|
{
|
|
GstVideoCodecState * const codec_state = decoder->codec_state;
|
|
|
|
if (codec_state->info.interlace_mode != mode) {
|
|
GST_DEBUG("interlace mode changed to %s",
|
|
gst_interlace_mode_to_string(mode));
|
|
codec_state->info.interlace_mode = mode;
|
|
gst_caps_set_simple(codec_state->caps, "interlaced",
|
|
G_TYPE_BOOLEAN, mode != GST_VIDEO_INTERLACE_MODE_PROGRESSIVE, NULL);
|
|
notify_codec_state_changed(decoder);
|
|
}
|
|
}
|
|
|
|
void
|
|
gst_vaapi_decoder_set_interlaced(GstVaapiDecoder *decoder, gboolean interlaced)
|
|
{
|
|
gst_vaapi_decoder_set_interlace_mode(decoder,
|
|
(interlaced ?
|
|
GST_VIDEO_INTERLACE_MODE_INTERLEAVED :
|
|
GST_VIDEO_INTERLACE_MODE_PROGRESSIVE));
|
|
}
|
|
|
|
gboolean
|
|
gst_vaapi_decoder_ensure_context(
|
|
GstVaapiDecoder *decoder,
|
|
GstVaapiContextInfo *cip
|
|
)
|
|
{
|
|
gst_vaapi_decoder_set_picture_size(decoder, cip->width, cip->height);
|
|
|
|
if (decoder->context) {
|
|
if (!gst_vaapi_context_reset_full(decoder->context, cip))
|
|
return FALSE;
|
|
}
|
|
else {
|
|
decoder->context = gst_vaapi_context_new_full(decoder->display, cip);
|
|
if (!decoder->context)
|
|
return FALSE;
|
|
}
|
|
decoder->va_context = gst_vaapi_context_get_id(decoder->context);
|
|
return TRUE;
|
|
}
|
|
|
|
void
|
|
gst_vaapi_decoder_push_frame(GstVaapiDecoder *decoder,
|
|
GstVideoCodecFrame *frame)
|
|
{
|
|
push_frame(decoder, frame);
|
|
}
|
|
|
|
GstVaapiDecoderStatus
|
|
gst_vaapi_decoder_check_status(GstVaapiDecoder *decoder)
|
|
{
|
|
if (decoder->context &&
|
|
gst_vaapi_context_get_surface_count(decoder->context) < 1)
|
|
return GST_VAAPI_DECODER_STATUS_ERROR_NO_SURFACE;
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
}
|
|
|
|
GstVaapiDecoderStatus
|
|
gst_vaapi_decoder_parse(GstVaapiDecoder *decoder,
|
|
GstVideoCodecFrame *base_frame, GstAdapter *adapter, gboolean at_eos,
|
|
guint *got_unit_size_ptr, gboolean *got_frame_ptr)
|
|
{
|
|
g_return_val_if_fail(decoder != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
g_return_val_if_fail(base_frame != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
g_return_val_if_fail(adapter != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
g_return_val_if_fail(got_unit_size_ptr != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
g_return_val_if_fail(got_frame_ptr != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
|
|
return do_parse(decoder, base_frame, adapter, at_eos,
|
|
got_unit_size_ptr, got_frame_ptr);
|
|
}
|
|
|
|
GstVaapiDecoderStatus
|
|
gst_vaapi_decoder_decode(GstVaapiDecoder *decoder, GstVideoCodecFrame *frame)
|
|
{
|
|
GstVaapiDecoderStatus status;
|
|
|
|
g_return_val_if_fail(decoder != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
g_return_val_if_fail(frame != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
g_return_val_if_fail(frame->user_data != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
|
|
status = gst_vaapi_decoder_check_status(decoder);
|
|
if (status != GST_VAAPI_DECODER_STATUS_SUCCESS)
|
|
return status;
|
|
return do_decode(decoder, frame);
|
|
}
|
|
|
|
GstVaapiDecoderStatus
|
|
gst_vaapi_decoder_flush(GstVaapiDecoder *decoder)
|
|
{
|
|
g_return_val_if_fail(decoder != NULL,
|
|
GST_VAAPI_DECODER_STATUS_ERROR_INVALID_PARAMETER);
|
|
|
|
return do_flush(decoder);
|
|
}
|
|
|
|
GstVaapiDecoderStatus
|
|
gst_vaapi_decoder_decode_codec_data(GstVaapiDecoder *decoder)
|
|
{
|
|
GstVaapiDecoderClass * const klass = GST_VAAPI_DECODER_GET_CLASS(decoder);
|
|
GstBuffer * const codec_data = GST_VAAPI_DECODER_CODEC_DATA(decoder);
|
|
GstVaapiDecoderStatus status;
|
|
GstMapInfo map_info;
|
|
const guchar *buf;
|
|
guint buf_size;
|
|
|
|
if (!codec_data)
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
|
|
/* FIXME: add a meaningful error code? */
|
|
if (!klass->decode_codec_data)
|
|
return GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
|
|
if (!gst_buffer_map(codec_data, &map_info, GST_MAP_READ)) {
|
|
GST_ERROR("failed to map buffer");
|
|
return GST_VAAPI_DECODER_STATUS_ERROR_UNKNOWN;
|
|
}
|
|
|
|
buf = map_info.data;
|
|
buf_size = map_info.size;
|
|
if (G_LIKELY(buf && buf_size > 0))
|
|
status = klass->decode_codec_data(decoder, buf, buf_size);
|
|
else
|
|
status = GST_VAAPI_DECODER_STATUS_SUCCESS;
|
|
gst_buffer_unmap(codec_data, &map_info);
|
|
return status;
|
|
}
|