mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-12-18 06:16:36 +00:00
a71cf4dfb2
The property limits the available output caps
1164 lines
32 KiB
C
1164 lines
32 KiB
C
/*
|
|
* GStreamer
|
|
* Copyright (C) 2009 Carl-Anton Ingmarsson <ca.ingmarsson@gmail.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
|
|
* Boston, MA 02111-1307, USA.
|
|
*/
|
|
|
|
/**
|
|
* SECTION:element-vdpaumpegdec
|
|
*
|
|
* FIXME:Describe vdpaumpegdec here.
|
|
*
|
|
* <refsect2>
|
|
* <title>Example launch line</title>
|
|
* |[
|
|
* gst-launch -v -m fakesrc ! vdpaumpegdec ! fakesink silent=TRUE
|
|
* ]|
|
|
* </refsect2>
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
# include <config.h>
|
|
#endif
|
|
|
|
#include <gst/gst.h>
|
|
#include <gst/base/gstbytereader.h>
|
|
#include <gst/base/gstbitreader.h>
|
|
#include <string.h>
|
|
|
|
#include "mpegutil.h"
|
|
#include "gstvdpvideosrcpad.h"
|
|
#include "gstvdpvideobuffer.h"
|
|
|
|
#include "gstvdpmpegdec.h"
|
|
|
|
GST_DEBUG_CATEGORY_STATIC (gst_vdp_mpeg_dec_debug);
|
|
#define GST_CAT_DEFAULT gst_vdp_mpeg_dec_debug
|
|
|
|
/* Filter signals and args */
|
|
enum
|
|
{
|
|
/* FILL ME */
|
|
LAST_SIGNAL
|
|
};
|
|
|
|
enum
|
|
{
|
|
PROP_0,
|
|
PROP_DISPLAY
|
|
};
|
|
|
|
/* the capabilities of the inputs and outputs.
|
|
*
|
|
* describe the real formats here.
|
|
*/
|
|
static GstStaticPadTemplate sink_template = GST_STATIC_PAD_TEMPLATE ("sink",
|
|
GST_PAD_SINK,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS ("video/mpeg, mpegversion = (int) [ 1, 2 ], "
|
|
"systemstream = (boolean) false, parsed = (boolean) true")
|
|
);
|
|
|
|
#define DEBUG_INIT(bla) \
|
|
GST_DEBUG_CATEGORY_INIT (gst_vdp_mpeg_dec_debug, "vdpaumpegdec", 0, "VDPAU powered mpeg decoder");
|
|
|
|
GST_BOILERPLATE_FULL (GstVdpMpegDec, gst_vdp_mpeg_dec,
|
|
GstElement, GST_TYPE_ELEMENT, DEBUG_INIT);
|
|
|
|
static void gst_vdp_mpeg_dec_init_info (VdpPictureInfoMPEG1Or2 * vdp_info);
|
|
|
|
typedef struct
|
|
{
|
|
GstByteReader reader;
|
|
GstBuffer *buffer;
|
|
guint start;
|
|
} GstVdpMpegPacketizer;
|
|
|
|
static GstBuffer *
|
|
gst_vdp_mpeg_packetizer_get_next_packet (GstVdpMpegPacketizer * packetizer)
|
|
{
|
|
guint offset, size;
|
|
GstBuffer *buf;
|
|
|
|
if (packetizer->start == -1)
|
|
return NULL;
|
|
|
|
if (!gst_byte_reader_set_pos (&packetizer->reader, packetizer->start + 3))
|
|
return NULL;
|
|
|
|
offset = gst_byte_reader_masked_scan_uint32 (&packetizer->reader, 0xffffff00,
|
|
0x00000100, 0, gst_byte_reader_get_remaining (&packetizer->reader));
|
|
|
|
if (offset != -1) {
|
|
offset = gst_byte_reader_get_pos (&packetizer->reader) + offset;
|
|
size = offset - packetizer->start;
|
|
} else
|
|
size = gst_byte_reader_get_remaining (&packetizer->reader) + 3;
|
|
|
|
buf = gst_buffer_create_sub (packetizer->buffer, packetizer->start, size);
|
|
|
|
packetizer->start = offset;
|
|
|
|
return buf;
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_packetizer_init (GstVdpMpegPacketizer * packetizer,
|
|
GstBuffer * buffer)
|
|
{
|
|
guint offset;
|
|
|
|
gst_byte_reader_init_from_buffer (&packetizer->reader, buffer);
|
|
packetizer->buffer = buffer;
|
|
|
|
offset = gst_byte_reader_masked_scan_uint32 (&packetizer->reader, 0xffffff00,
|
|
0x00000100, 0, gst_byte_reader_get_remaining (&packetizer->reader));
|
|
|
|
packetizer->start = offset;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_set_caps (GstPad * pad, GstCaps * caps)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec = GST_VDP_MPEG_DEC (gst_pad_get_parent (pad));
|
|
GstStructure *structure;
|
|
|
|
gint width, height;
|
|
gint fps_n, fps_d;
|
|
gint par_n, par_d;
|
|
gboolean interlaced = FALSE;
|
|
|
|
GstCaps *src_caps;
|
|
gboolean res;
|
|
|
|
const GValue *value;
|
|
|
|
structure = gst_caps_get_structure (caps, 0);
|
|
|
|
/* create src_pad caps */
|
|
gst_structure_get_int (structure, "width", &width);
|
|
gst_structure_get_int (structure, "height", &height);
|
|
gst_structure_get_fraction (structure, "framerate", &fps_n, &fps_d);
|
|
gst_structure_get_fraction (structure, "pixel-aspect-ratio", &par_n, &par_d);
|
|
gst_structure_get_boolean (structure, "interlaced", &interlaced);
|
|
|
|
src_caps = gst_pad_get_allowed_caps (mpeg_dec->src);
|
|
if (!src_caps)
|
|
goto error;
|
|
if (gst_caps_is_empty (src_caps))
|
|
goto error;
|
|
|
|
gst_caps_truncate (src_caps);
|
|
gst_caps_set_simple (src_caps,
|
|
"width", G_TYPE_INT, width,
|
|
"height", G_TYPE_INT, height,
|
|
"framerate", GST_TYPE_FRACTION, fps_n, fps_d,
|
|
"pixel-aspect-ratio", GST_TYPE_FRACTION, par_n, par_d,
|
|
"interlaced", G_TYPE_BOOLEAN, interlaced, NULL);
|
|
gst_pad_fixate_caps (mpeg_dec->src, src_caps);
|
|
|
|
structure = gst_caps_get_structure (src_caps, 0);
|
|
|
|
GST_DEBUG_OBJECT (mpeg_dec, "Setting source caps to %" GST_PTR_FORMAT,
|
|
src_caps);
|
|
|
|
res = gst_vdp_video_src_pad_set_caps
|
|
(GST_VDP_VIDEO_SRC_PAD (mpeg_dec->src), src_caps);
|
|
gst_caps_unref (src_caps);
|
|
if (!res)
|
|
goto done;
|
|
|
|
mpeg_dec->width = width;
|
|
mpeg_dec->height = height;
|
|
mpeg_dec->fps_n = fps_n;
|
|
mpeg_dec->fps_d = fps_d;
|
|
mpeg_dec->interlaced = interlaced;
|
|
|
|
structure = gst_caps_get_structure (caps, 0);
|
|
/* parse caps to setup decoder */
|
|
gst_structure_get_int (structure, "mpegversion", &mpeg_dec->version);
|
|
|
|
/* Default to MPEG1 until we find otherwise */
|
|
mpeg_dec->profile = VDP_DECODER_PROFILE_MPEG1;
|
|
|
|
value = gst_structure_get_value (structure, "codec_data");
|
|
if (value) {
|
|
GstBuffer *codec_data, *buf;
|
|
GstVdpMpegPacketizer packetizer;
|
|
|
|
codec_data = gst_value_get_buffer (value);
|
|
gst_vdp_mpeg_packetizer_init (&packetizer, codec_data);
|
|
if ((buf = gst_vdp_mpeg_packetizer_get_next_packet (&packetizer))) {
|
|
MPEGSeqHdr hdr;
|
|
guint32 bitrate;
|
|
|
|
mpeg_util_parse_sequence_hdr (&hdr, buf);
|
|
|
|
memcpy (&mpeg_dec->vdp_info.intra_quantizer_matrix,
|
|
&hdr.intra_quantizer_matrix, 64);
|
|
memcpy (&mpeg_dec->vdp_info.non_intra_quantizer_matrix,
|
|
&hdr.non_intra_quantizer_matrix, 64);
|
|
|
|
bitrate = hdr.bitrate;
|
|
gst_buffer_unref (buf);
|
|
|
|
if ((buf = gst_vdp_mpeg_packetizer_get_next_packet (&packetizer))) {
|
|
MPEGSeqExtHdr ext;
|
|
|
|
mpeg_util_parse_sequence_extension (&ext, buf);
|
|
if (mpeg_dec->version != 1) {
|
|
switch (ext.profile) {
|
|
case 5:
|
|
mpeg_dec->profile = VDP_DECODER_PROFILE_MPEG2_SIMPLE;
|
|
break;
|
|
default:
|
|
mpeg_dec->profile = VDP_DECODER_PROFILE_MPEG2_MAIN;
|
|
break;
|
|
}
|
|
}
|
|
|
|
bitrate += (ext.bitrate_ext << 18);
|
|
gst_buffer_unref (buf);
|
|
}
|
|
|
|
mpeg_dec->duration =
|
|
gst_util_uint64_scale (1, GST_SECOND * mpeg_dec->fps_d,
|
|
mpeg_dec->fps_n);
|
|
|
|
mpeg_dec->byterate = bitrate * 50;
|
|
GST_DEBUG ("byterate: %" G_GINT64_FORMAT, mpeg_dec->byterate);
|
|
}
|
|
}
|
|
|
|
res = TRUE;
|
|
|
|
done:
|
|
gst_object_unref (mpeg_dec);
|
|
|
|
return res;
|
|
|
|
error:
|
|
res = FALSE;
|
|
goto done;
|
|
}
|
|
|
|
GstFlowReturn
|
|
gst_vdp_mpeg_dec_push_video_buffer (GstVdpMpegDec * mpeg_dec,
|
|
GstVdpVideoBuffer * buf)
|
|
{
|
|
gint64 byterate;
|
|
|
|
if (GST_BUFFER_TIMESTAMP (buf) == GST_CLOCK_TIME_NONE
|
|
&& GST_CLOCK_TIME_IS_VALID (mpeg_dec->next_timestamp)) {
|
|
GST_BUFFER_TIMESTAMP (buf) = mpeg_dec->next_timestamp;
|
|
} else if (GST_BUFFER_TIMESTAMP (buf) == GST_CLOCK_TIME_NONE) {
|
|
GST_BUFFER_TIMESTAMP (buf) = gst_util_uint64_scale (mpeg_dec->frame_nr,
|
|
GST_SECOND * mpeg_dec->fps_d, mpeg_dec->fps_n);
|
|
}
|
|
|
|
if (mpeg_dec->seeking) {
|
|
GstEvent *event;
|
|
|
|
event = gst_event_new_new_segment (FALSE,
|
|
mpeg_dec->segment.rate, GST_FORMAT_TIME, GST_BUFFER_TIMESTAMP (buf),
|
|
mpeg_dec->segment.stop, GST_BUFFER_TIMESTAMP (buf));
|
|
|
|
gst_pad_push_event (mpeg_dec->src, event);
|
|
|
|
mpeg_dec->seeking = FALSE;
|
|
}
|
|
|
|
mpeg_dec->next_timestamp = GST_BUFFER_TIMESTAMP (buf) +
|
|
GST_BUFFER_DURATION (buf);
|
|
|
|
gst_segment_set_last_stop (&mpeg_dec->segment, GST_FORMAT_TIME,
|
|
GST_BUFFER_TIMESTAMP (buf));
|
|
|
|
mpeg_dec->accumulated_duration += GST_BUFFER_DURATION (buf);
|
|
mpeg_dec->accumulated_size += GST_BUFFER_SIZE (buf);
|
|
if (mpeg_dec->accumulated_duration && mpeg_dec->accumulated_size != 0) {
|
|
byterate = gst_util_uint64_scale (mpeg_dec->accumulated_size, GST_SECOND,
|
|
mpeg_dec->accumulated_duration);
|
|
GST_DEBUG ("byterate: %" G_GINT64_FORMAT, mpeg_dec->byterate);
|
|
|
|
mpeg_dec->byterate = (mpeg_dec->byterate + byterate) / 2;
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (mpeg_dec,
|
|
"Pushing buffer with timestamp: %" GST_TIME_FORMAT
|
|
" frame_nr: %" G_GINT64_FORMAT,
|
|
GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (buf)), GST_BUFFER_OFFSET (buf));
|
|
|
|
return gst_vdp_video_src_pad_push ((GstVdpVideoSrcPad *) mpeg_dec->src, buf);
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_vdp_mpeg_dec_alloc_buffer (GstVdpMpegDec * mpeg_dec,
|
|
GstVdpVideoBuffer ** outbuf)
|
|
{
|
|
GstVdpVideoSrcPad *vdp_pad;
|
|
GstFlowReturn ret = GST_FLOW_OK;
|
|
|
|
vdp_pad = (GstVdpVideoSrcPad *) mpeg_dec->src;
|
|
ret = gst_vdp_video_src_pad_alloc_buffer (vdp_pad, outbuf);
|
|
if (ret != GST_FLOW_OK)
|
|
return ret;
|
|
|
|
if (mpeg_dec->decoder == VDP_INVALID_HANDLE) {
|
|
GstVdpDevice *device;
|
|
VdpStatus status;
|
|
|
|
device = GST_VDP_VIDEO_BUFFER (*outbuf)->device;
|
|
|
|
status = device->vdp_decoder_create (device->device, mpeg_dec->profile,
|
|
mpeg_dec->width, mpeg_dec->height, 2, &mpeg_dec->decoder);
|
|
if (status != VDP_STATUS_OK) {
|
|
GST_ELEMENT_ERROR (mpeg_dec, RESOURCE, READ,
|
|
("Could not create vdpau decoder"),
|
|
("Error returned from vdpau was: %s",
|
|
device->vdp_get_error_string (status)));
|
|
gst_buffer_unref (GST_BUFFER_CAST (*outbuf));
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
}
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_vdp_mpeg_dec_decode (GstVdpMpegDec * mpeg_dec,
|
|
GstClockTime timestamp, gint64 size)
|
|
{
|
|
VdpPictureInfoMPEG1Or2 *info;
|
|
GstBuffer *buffer;
|
|
GstVdpVideoBuffer *outbuf;
|
|
VdpVideoSurface surface;
|
|
GstVdpDevice *device;
|
|
VdpBitstreamBuffer vbit[1];
|
|
VdpStatus status;
|
|
|
|
info = &mpeg_dec->vdp_info;
|
|
|
|
if (info->picture_coding_type != B_FRAME) {
|
|
if (info->backward_reference != VDP_INVALID_HANDLE) {
|
|
gst_buffer_ref (GST_BUFFER_CAST (mpeg_dec->b_buffer));
|
|
gst_vdp_mpeg_dec_push_video_buffer (mpeg_dec, mpeg_dec->b_buffer);
|
|
}
|
|
|
|
if (info->forward_reference != VDP_INVALID_HANDLE) {
|
|
gst_buffer_unref (GST_BUFFER_CAST (mpeg_dec->f_buffer));
|
|
info->forward_reference = VDP_INVALID_HANDLE;
|
|
}
|
|
|
|
info->forward_reference = info->backward_reference;
|
|
mpeg_dec->f_buffer = mpeg_dec->b_buffer;
|
|
|
|
info->backward_reference = VDP_INVALID_HANDLE;
|
|
}
|
|
|
|
if (gst_vdp_mpeg_dec_alloc_buffer (mpeg_dec, &outbuf) != GST_FLOW_OK) {
|
|
gst_adapter_clear (mpeg_dec->adapter);
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
device = GST_VDP_VIDEO_BUFFER (outbuf)->device;
|
|
|
|
GST_BUFFER_TIMESTAMP (outbuf) = timestamp;
|
|
GST_BUFFER_DURATION (outbuf) = mpeg_dec->duration;
|
|
GST_BUFFER_OFFSET (outbuf) = mpeg_dec->frame_nr;
|
|
GST_BUFFER_SIZE (outbuf) = size;
|
|
|
|
if (info->picture_coding_type == I_FRAME)
|
|
GST_BUFFER_FLAG_UNSET (outbuf, GST_BUFFER_FLAG_DELTA_UNIT);
|
|
else
|
|
GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_DELTA_UNIT);
|
|
|
|
if (info->top_field_first)
|
|
GST_BUFFER_FLAG_SET (outbuf, GST_VIDEO_BUFFER_TFF);
|
|
else
|
|
GST_BUFFER_FLAG_UNSET (outbuf, GST_VIDEO_BUFFER_TFF);
|
|
|
|
buffer = gst_adapter_take_buffer (mpeg_dec->adapter,
|
|
gst_adapter_available (mpeg_dec->adapter));
|
|
|
|
surface = GST_VDP_VIDEO_BUFFER (outbuf)->surface;
|
|
|
|
vbit[0].struct_version = VDP_BITSTREAM_BUFFER_VERSION;
|
|
vbit[0].bitstream = GST_BUFFER_DATA (buffer);
|
|
vbit[0].bitstream_bytes = GST_BUFFER_SIZE (buffer);
|
|
|
|
status = device->vdp_decoder_render (mpeg_dec->decoder, surface,
|
|
(VdpPictureInfo *) info, 1, vbit);
|
|
gst_buffer_unref (buffer);
|
|
info->slice_count = 0;
|
|
|
|
if (status != VDP_STATUS_OK) {
|
|
GST_ELEMENT_ERROR (mpeg_dec, RESOURCE, READ,
|
|
("Could not decode"),
|
|
("Error returned from vdpau was: %s",
|
|
device->vdp_get_error_string (status)));
|
|
|
|
gst_buffer_unref (GST_BUFFER_CAST (outbuf));
|
|
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
if (info->picture_coding_type == B_FRAME) {
|
|
gst_vdp_mpeg_dec_push_video_buffer (mpeg_dec, outbuf);
|
|
} else {
|
|
info->backward_reference = surface;
|
|
mpeg_dec->b_buffer = outbuf;
|
|
}
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_parse_picture_coding (GstVdpMpegDec * mpeg_dec,
|
|
GstBuffer * buffer)
|
|
{
|
|
MPEGPictureExt pic_ext;
|
|
VdpPictureInfoMPEG1Or2 *info;
|
|
gint fields;
|
|
|
|
info = &mpeg_dec->vdp_info;
|
|
|
|
if (!mpeg_util_parse_picture_coding_extension (&pic_ext, buffer))
|
|
return FALSE;
|
|
|
|
memcpy (&mpeg_dec->vdp_info.f_code, &pic_ext.f_code, 4);
|
|
|
|
info->intra_dc_precision = pic_ext.intra_dc_precision;
|
|
info->picture_structure = pic_ext.picture_structure;
|
|
info->top_field_first = pic_ext.top_field_first;
|
|
info->frame_pred_frame_dct = pic_ext.frame_pred_frame_dct;
|
|
info->concealment_motion_vectors = pic_ext.concealment_motion_vectors;
|
|
info->q_scale_type = pic_ext.q_scale_type;
|
|
info->intra_vlc_format = pic_ext.intra_vlc_format;
|
|
info->alternate_scan = pic_ext.alternate_scan;
|
|
|
|
fields = 2;
|
|
if (pic_ext.picture_structure == 3) {
|
|
if (mpeg_dec->interlaced) {
|
|
if (pic_ext.progressive_frame == 0)
|
|
fields = 2;
|
|
if (pic_ext.progressive_frame == 0 && pic_ext.repeat_first_field == 0)
|
|
fields = 2;
|
|
if (pic_ext.progressive_frame == 1 && pic_ext.repeat_first_field == 1)
|
|
fields = 3;
|
|
} else {
|
|
if (pic_ext.repeat_first_field == 0)
|
|
fields = 2;
|
|
if (pic_ext.repeat_first_field == 1 && pic_ext.top_field_first == 0)
|
|
fields = 4;
|
|
if (pic_ext.repeat_first_field == 1 && pic_ext.top_field_first == 1)
|
|
fields = 6;
|
|
}
|
|
} else
|
|
fields = 1;
|
|
|
|
GST_DEBUG ("fields: %d", fields);
|
|
|
|
mpeg_dec->duration = gst_util_uint64_scale (fields,
|
|
GST_SECOND * mpeg_dec->fps_d, 2 * mpeg_dec->fps_n);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_parse_sequence (GstVdpMpegDec * mpeg_dec, GstBuffer * buffer)
|
|
{
|
|
MPEGSeqHdr hdr;
|
|
|
|
if (!mpeg_util_parse_sequence_hdr (&hdr, buffer))
|
|
return FALSE;
|
|
|
|
memcpy (&mpeg_dec->vdp_info.intra_quantizer_matrix,
|
|
&hdr.intra_quantizer_matrix, 64);
|
|
memcpy (&mpeg_dec->vdp_info.non_intra_quantizer_matrix,
|
|
&hdr.non_intra_quantizer_matrix, 64);
|
|
|
|
if (mpeg_dec->state == GST_VDP_MPEG_DEC_NEED_SEQUENCE)
|
|
mpeg_dec->state = GST_VDP_MPEG_DEC_NEED_DATA;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_parse_picture (GstVdpMpegDec * mpeg_dec, GstBuffer * buffer)
|
|
{
|
|
MPEGPictureHdr pic_hdr;
|
|
|
|
if (!mpeg_util_parse_picture_hdr (&pic_hdr, buffer))
|
|
return FALSE;
|
|
|
|
if (pic_hdr.pic_type != I_FRAME
|
|
&& mpeg_dec->vdp_info.backward_reference == VDP_INVALID_HANDLE) {
|
|
GST_DEBUG_OBJECT (mpeg_dec,
|
|
"Drop frame since we haven't got an I_FRAME yet");
|
|
return FALSE;
|
|
}
|
|
if (pic_hdr.pic_type == B_FRAME
|
|
&& mpeg_dec->vdp_info.forward_reference == VDP_INVALID_HANDLE) {
|
|
GST_DEBUG_OBJECT (mpeg_dec,
|
|
"Drop frame since we haven't got two non B_FRAMES yet");
|
|
return FALSE;
|
|
}
|
|
|
|
mpeg_dec->vdp_info.picture_coding_type = pic_hdr.pic_type;
|
|
|
|
if (mpeg_dec->version == 1) {
|
|
mpeg_dec->vdp_info.full_pel_forward_vector =
|
|
pic_hdr.full_pel_forward_vector;
|
|
mpeg_dec->vdp_info.full_pel_backward_vector =
|
|
pic_hdr.full_pel_backward_vector;
|
|
memcpy (&mpeg_dec->vdp_info.f_code, &pic_hdr.f_code, 4);
|
|
}
|
|
|
|
mpeg_dec->frame_nr = mpeg_dec->gop_frame + pic_hdr.tsn;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_parse_gop (GstVdpMpegDec * mpeg_dec, GstBuffer * buffer)
|
|
{
|
|
MPEGGop gop;
|
|
GstClockTime time;
|
|
|
|
if (!mpeg_util_parse_gop (&gop, buffer))
|
|
return FALSE;
|
|
|
|
time = GST_SECOND * (gop.hour * 3600 + gop.minute * 60 + gop.second);
|
|
|
|
GST_DEBUG ("gop timestamp: %" GST_TIME_FORMAT, GST_TIME_ARGS (time));
|
|
|
|
mpeg_dec->gop_frame =
|
|
gst_util_uint64_scale (time, mpeg_dec->fps_n,
|
|
mpeg_dec->fps_d * GST_SECOND) + gop.frame;
|
|
|
|
if (mpeg_dec->state == GST_VDP_MPEG_DEC_NEED_GOP)
|
|
mpeg_dec->state = GST_VDP_MPEG_DEC_NEED_DATA;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_parse_quant_matrix (GstVdpMpegDec * mpeg_dec,
|
|
GstBuffer * buffer)
|
|
{
|
|
MPEGQuantMatrix qm;
|
|
|
|
if (!mpeg_util_parse_quant_matrix (&qm, buffer))
|
|
return FALSE;
|
|
|
|
memcpy (&mpeg_dec->vdp_info.intra_quantizer_matrix,
|
|
&qm.intra_quantizer_matrix, 64);
|
|
memcpy (&mpeg_dec->vdp_info.non_intra_quantizer_matrix,
|
|
&qm.non_intra_quantizer_matrix, 64);
|
|
return TRUE;
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_dec_flush (GstVdpMpegDec * mpeg_dec)
|
|
{
|
|
if (mpeg_dec->vdp_info.forward_reference != VDP_INVALID_HANDLE)
|
|
gst_buffer_unref (GST_BUFFER_CAST (mpeg_dec->f_buffer));
|
|
if (mpeg_dec->vdp_info.backward_reference != VDP_INVALID_HANDLE)
|
|
gst_buffer_unref (GST_BUFFER_CAST (mpeg_dec->b_buffer));
|
|
|
|
gst_adapter_clear (mpeg_dec->adapter);
|
|
|
|
gst_vdp_mpeg_dec_init_info (&mpeg_dec->vdp_info);
|
|
mpeg_dec->next_timestamp = GST_CLOCK_TIME_NONE;
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_dec_start (GstVdpMpegDec * mpeg_dec)
|
|
{
|
|
gst_vdp_mpeg_dec_init_info (&mpeg_dec->vdp_info);
|
|
|
|
mpeg_dec->decoder = VDP_INVALID_HANDLE;
|
|
|
|
mpeg_dec->state = GST_VDP_MPEG_DEC_NEED_SEQUENCE;
|
|
|
|
gst_segment_init (&mpeg_dec->segment, GST_FORMAT_TIME);
|
|
mpeg_dec->seeking = FALSE;
|
|
|
|
mpeg_dec->accumulated_size = 0;
|
|
mpeg_dec->accumulated_duration = 0;
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_dec_stop (GstVdpMpegDec * mpeg_dec)
|
|
{
|
|
GstVdpDevice *device;
|
|
|
|
if ((device =
|
|
gst_vdp_video_src_pad_get_device (GST_VDP_VIDEO_SRC_PAD
|
|
(mpeg_dec->src)))) {
|
|
if (mpeg_dec->decoder != VDP_INVALID_HANDLE)
|
|
device->vdp_decoder_destroy (mpeg_dec->decoder);
|
|
}
|
|
|
|
if (mpeg_dec->vdp_info.forward_reference != VDP_INVALID_HANDLE)
|
|
mpeg_dec->vdp_info.forward_reference = VDP_INVALID_HANDLE;
|
|
if (mpeg_dec->vdp_info.backward_reference != VDP_INVALID_HANDLE)
|
|
mpeg_dec->vdp_info.backward_reference = VDP_INVALID_HANDLE;
|
|
|
|
gst_adapter_clear (mpeg_dec->adapter);
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_vdp_mpeg_dec_chain (GstPad * pad, GstBuffer * buffer)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec = GST_VDP_MPEG_DEC (gst_pad_get_parent (pad));
|
|
GstVdpMpegPacketizer packetizer;
|
|
GstBuffer *buf;
|
|
GstFlowReturn ret = GST_FLOW_OK;
|
|
|
|
if (G_UNLIKELY (GST_BUFFER_FLAG_IS_SET (buffer, GST_BUFFER_FLAG_DISCONT))) {
|
|
GST_DEBUG_OBJECT (mpeg_dec, "Received discont buffer");
|
|
gst_vdp_mpeg_dec_flush (mpeg_dec);
|
|
}
|
|
|
|
|
|
gst_vdp_mpeg_packetizer_init (&packetizer, buffer);
|
|
while ((buf = gst_vdp_mpeg_packetizer_get_next_packet (&packetizer))) {
|
|
GstBitReader b_reader = GST_BIT_READER_INIT_FROM_BUFFER (buf);
|
|
guint32 sync_code;
|
|
guint8 start_code;
|
|
|
|
/* skip sync_code */
|
|
gst_bit_reader_get_bits_uint32 (&b_reader, &sync_code, 8 * 3);
|
|
|
|
/* start_code */
|
|
gst_bit_reader_get_bits_uint8 (&b_reader, &start_code, 8);
|
|
|
|
if (start_code >= MPEG_PACKET_SLICE_MIN
|
|
&& start_code <= MPEG_PACKET_SLICE_MAX) {
|
|
GST_DEBUG_OBJECT (mpeg_dec, "MPEG_PACKET_SLICE");
|
|
|
|
gst_buffer_ref (buf);
|
|
gst_adapter_push (mpeg_dec->adapter, buf);
|
|
mpeg_dec->vdp_info.slice_count++;
|
|
}
|
|
|
|
switch (start_code) {
|
|
case MPEG_PACKET_PICTURE:
|
|
GST_DEBUG_OBJECT (mpeg_dec, "MPEG_PACKET_PICTURE");
|
|
|
|
if (!gst_vdp_mpeg_dec_parse_picture (mpeg_dec, buf))
|
|
goto done;
|
|
|
|
break;
|
|
case MPEG_PACKET_SEQUENCE:
|
|
GST_DEBUG_OBJECT (mpeg_dec, "MPEG_PACKET_SEQUENCE");
|
|
gst_vdp_mpeg_dec_parse_sequence (mpeg_dec, buf);
|
|
break;
|
|
case MPEG_PACKET_EXTENSION:
|
|
{
|
|
guint8 ext_code;
|
|
|
|
GST_DEBUG_OBJECT (mpeg_dec, "MPEG_PACKET_EXTENSION");
|
|
|
|
/* ext_code */
|
|
gst_bit_reader_get_bits_uint8 (&b_reader, &ext_code, 4);
|
|
switch (ext_code) {
|
|
case MPEG_PACKET_EXT_PICTURE_CODING:
|
|
GST_DEBUG_OBJECT (mpeg_dec, "MPEG_PACKET_EXT_PICTURE_CODING");
|
|
gst_vdp_mpeg_dec_parse_picture_coding (mpeg_dec, buf);
|
|
break;
|
|
case MPEG_PACKET_EXT_QUANT_MATRIX:
|
|
GST_DEBUG_OBJECT (mpeg_dec, "MPEG_PACKET_EXT_QUANT_MATRIX");
|
|
gst_vdp_mpeg_dec_parse_quant_matrix (mpeg_dec, buf);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
case MPEG_PACKET_GOP:
|
|
GST_DEBUG_OBJECT (mpeg_dec, "MPEG_PACKET_GOP");
|
|
gst_vdp_mpeg_dec_parse_gop (mpeg_dec, buf);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
gst_buffer_unref (buf);
|
|
}
|
|
|
|
if (mpeg_dec->state == GST_VDP_MPEG_DEC_NEED_SEQUENCE ||
|
|
mpeg_dec->state == GST_VDP_MPEG_DEC_NEED_GOP) {
|
|
gst_adapter_clear (mpeg_dec->adapter);
|
|
goto done;
|
|
}
|
|
|
|
if (mpeg_dec->vdp_info.slice_count > 0)
|
|
ret = gst_vdp_mpeg_dec_decode (mpeg_dec, GST_BUFFER_TIMESTAMP (buffer),
|
|
GST_BUFFER_SIZE (buffer));
|
|
|
|
done:
|
|
gst_object_unref (mpeg_dec);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_convert (GstVdpMpegDec * mpeg_dec,
|
|
GstFormat src_format, gint64 src_value,
|
|
GstFormat dest_format, gint64 * dest_value)
|
|
{
|
|
|
|
if (src_format == dest_format) {
|
|
*dest_value = src_value;
|
|
return TRUE;
|
|
}
|
|
|
|
if (mpeg_dec->byterate == -1)
|
|
return FALSE;
|
|
|
|
if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_TIME) {
|
|
*dest_value = gst_util_uint64_scale (GST_SECOND, src_value,
|
|
mpeg_dec->byterate);
|
|
return TRUE;
|
|
}
|
|
|
|
if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_BYTES) {
|
|
*dest_value =
|
|
gst_util_uint64_scale_int (src_value, mpeg_dec->byterate, GST_SECOND);
|
|
return TRUE;
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static const GstQueryType *
|
|
gst_mpeg_dec_get_querytypes (GstPad * pad)
|
|
{
|
|
static const GstQueryType list[] = {
|
|
GST_QUERY_POSITION,
|
|
GST_QUERY_DURATION,
|
|
0
|
|
};
|
|
|
|
return list;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_src_query (GstPad * pad, GstQuery * query)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec = GST_VDP_MPEG_DEC (gst_pad_get_parent (pad));
|
|
gboolean res;
|
|
|
|
switch (GST_QUERY_TYPE (query)) {
|
|
case GST_QUERY_POSITION:
|
|
{
|
|
GstFormat format;
|
|
|
|
if ((res = gst_pad_query_default (pad, query)))
|
|
goto done;
|
|
|
|
gst_query_parse_position (query, &format, NULL);
|
|
if (format == GST_FORMAT_TIME &&
|
|
GST_CLOCK_TIME_IS_VALID (mpeg_dec->next_timestamp)) {
|
|
gst_query_set_position (query, GST_FORMAT_TIME,
|
|
mpeg_dec->next_timestamp);
|
|
res = TRUE;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case GST_QUERY_DURATION:
|
|
{
|
|
GstFormat format;
|
|
|
|
if ((res = gst_pad_query_default (pad, query)))
|
|
goto done;
|
|
|
|
gst_query_parse_duration (query, &format, NULL);
|
|
if (format == GST_FORMAT_TIME) {
|
|
gint64 bytes;
|
|
|
|
format = GST_FORMAT_BYTES;
|
|
if (gst_pad_query_duration (pad, &format, &bytes)
|
|
&& format == GST_FORMAT_BYTES) {
|
|
gint64 duration;
|
|
|
|
if (gst_vdp_mpeg_dec_convert (mpeg_dec, GST_FORMAT_BYTES,
|
|
bytes, GST_FORMAT_TIME, &duration)) {
|
|
GST_DEBUG ("duration: %" GST_TIME_FORMAT, GST_TIME_ARGS (duration));
|
|
gst_query_set_duration (query, GST_FORMAT_TIME, duration);
|
|
res = TRUE;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
default:
|
|
res = gst_pad_query_default (pad, query);
|
|
}
|
|
|
|
done:
|
|
gst_object_unref (mpeg_dec);
|
|
|
|
return res;
|
|
}
|
|
|
|
static gboolean
|
|
normal_seek (GstVdpMpegDec * mpeg_dec, GstEvent * event)
|
|
{
|
|
gdouble rate;
|
|
GstFormat format;
|
|
GstSeekFlags flags;
|
|
GstSeekType cur_type, stop_type;
|
|
gint64 time_cur, bytes_cur;
|
|
gint64 time_stop, bytes_stop;
|
|
gboolean res;
|
|
gboolean update;
|
|
GstEvent *peer_event;
|
|
|
|
GST_DEBUG ("normal seek");
|
|
|
|
gst_event_parse_seek (event, &rate, &format, &flags,
|
|
&cur_type, &time_cur, &stop_type, &time_stop);
|
|
|
|
if (format != GST_FORMAT_TIME)
|
|
return FALSE;
|
|
|
|
gst_segment_set_seek (&mpeg_dec->segment, rate, GST_FORMAT_TIME, flags,
|
|
cur_type, time_cur, stop_type, time_stop, &update);
|
|
|
|
if (update) {
|
|
/* seek on bytes */
|
|
if (!gst_vdp_mpeg_dec_convert (mpeg_dec, GST_FORMAT_TIME, time_cur,
|
|
GST_FORMAT_BYTES, &bytes_cur))
|
|
goto convert_failed;
|
|
if (!gst_vdp_mpeg_dec_convert (mpeg_dec, GST_FORMAT_TIME, time_stop,
|
|
GST_FORMAT_BYTES, &bytes_stop))
|
|
goto convert_failed;
|
|
|
|
/* conversion succeeded, create the seek */
|
|
peer_event =
|
|
gst_event_new_seek (rate, GST_FORMAT_BYTES, flags,
|
|
cur_type, bytes_cur, stop_type, bytes_stop);
|
|
|
|
g_mutex_lock (mpeg_dec->mutex);
|
|
|
|
/* do the seek */
|
|
res = gst_pad_push_event (mpeg_dec->sink, peer_event);
|
|
|
|
if (res) {
|
|
mpeg_dec->state = GST_VDP_MPEG_DEC_NEED_GOP;
|
|
mpeg_dec->seeking = TRUE;
|
|
}
|
|
|
|
g_mutex_unlock (mpeg_dec->mutex);
|
|
|
|
} else {
|
|
GstEvent *event;
|
|
|
|
/* send segment with new rate */
|
|
event = gst_event_new_new_segment (TRUE,
|
|
mpeg_dec->segment.rate, GST_FORMAT_TIME, mpeg_dec->segment.start,
|
|
mpeg_dec->segment.stop, mpeg_dec->segment.time);
|
|
|
|
gst_pad_push_event (mpeg_dec->src, event);
|
|
res = TRUE;
|
|
}
|
|
|
|
return res;
|
|
|
|
/* ERRORS */
|
|
convert_failed:
|
|
{
|
|
/* probably unsupported seek format */
|
|
GST_DEBUG_OBJECT (mpeg_dec,
|
|
"failed to convert format %u into GST_FORMAT_TIME", format);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_src_event (GstPad * pad, GstEvent * event)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec = GST_VDP_MPEG_DEC (gst_pad_get_parent (pad));
|
|
gboolean res;
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_SEEK:
|
|
{
|
|
if ((res = gst_pad_event_default (pad, event)))
|
|
goto done;
|
|
|
|
res = normal_seek (mpeg_dec, event);
|
|
|
|
break;
|
|
}
|
|
default:
|
|
res = gst_pad_event_default (pad, event);
|
|
}
|
|
|
|
done:
|
|
gst_object_unref (mpeg_dec);
|
|
|
|
return res;
|
|
}
|
|
|
|
static gboolean
|
|
gst_vdp_mpeg_dec_sink_event (GstPad * pad, GstEvent * event)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec = GST_VDP_MPEG_DEC (gst_pad_get_parent (pad));
|
|
gboolean res;
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_FLUSH_STOP:
|
|
{
|
|
GST_DEBUG_OBJECT (mpeg_dec, "flush stop");
|
|
|
|
gst_vdp_mpeg_dec_flush (mpeg_dec);
|
|
res = gst_pad_push_event (mpeg_dec->src, event);
|
|
|
|
break;
|
|
}
|
|
case GST_EVENT_NEWSEGMENT:
|
|
{
|
|
gboolean update;
|
|
gdouble rate;
|
|
GstFormat format;
|
|
gint64 start;
|
|
gint64 stop;
|
|
gint64 position;
|
|
|
|
gst_event_parse_new_segment (event, &update, &rate, &format,
|
|
&start, &stop, &position);
|
|
|
|
if (format != GST_FORMAT_TIME) {
|
|
if (!gst_vdp_mpeg_dec_convert (mpeg_dec, format, start,
|
|
GST_FORMAT_TIME, &start))
|
|
goto convert_error;
|
|
if (!gst_vdp_mpeg_dec_convert (mpeg_dec, format, stop,
|
|
GST_FORMAT_TIME, &stop))
|
|
goto convert_error;
|
|
if (!gst_vdp_mpeg_dec_convert (mpeg_dec, format, position,
|
|
GST_FORMAT_TIME, &position))
|
|
goto convert_error;
|
|
|
|
gst_event_unref (event);
|
|
event = gst_event_new_new_segment (update, rate, GST_FORMAT_TIME, start,
|
|
stop, position);
|
|
}
|
|
|
|
g_mutex_lock (mpeg_dec->mutex);
|
|
/* if we seek ourselves we don't push out a newsegment now since we
|
|
* use the calculated timestamp of the first frame for this */
|
|
if (mpeg_dec->seeking) {
|
|
gst_event_unref (event);
|
|
res = TRUE;
|
|
g_mutex_unlock (mpeg_dec->mutex);
|
|
goto done;
|
|
}
|
|
g_mutex_unlock (mpeg_dec->mutex);
|
|
|
|
GST_DEBUG_OBJECT (mpeg_dec,
|
|
"Pushing new segment update %d format %d start %"
|
|
GST_TIME_FORMAT " stop %" GST_TIME_FORMAT " position %"
|
|
GST_TIME_FORMAT, update, format, GST_TIME_ARGS (start),
|
|
GST_TIME_ARGS (stop), GST_TIME_ARGS (position));
|
|
convert_error:
|
|
res = gst_pad_push_event (mpeg_dec->src, event);
|
|
|
|
break;
|
|
}
|
|
default:
|
|
res = gst_pad_event_default (pad, event);
|
|
}
|
|
|
|
done:
|
|
gst_object_unref (mpeg_dec);
|
|
|
|
return res;
|
|
}
|
|
|
|
static GstStateChangeReturn
|
|
gst_vdp_mpeg_dec_change_state (GstElement * element, GstStateChange transition)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec;
|
|
GstStateChangeReturn ret;
|
|
|
|
mpeg_dec = GST_VDP_MPEG_DEC (element);
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_READY_TO_PAUSED:
|
|
gst_vdp_mpeg_dec_start (mpeg_dec);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_PAUSED_TO_READY:
|
|
gst_vdp_mpeg_dec_stop (mpeg_dec);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* GObject vmethod implementations */
|
|
static void
|
|
gst_vdp_mpeg_dec_finalize (GObject * object)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec = (GstVdpMpegDec *) object;
|
|
|
|
g_object_unref (mpeg_dec->adapter);
|
|
g_mutex_free (mpeg_dec->mutex);
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_dec_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec = (GstVdpMpegDec *) object;
|
|
|
|
switch (prop_id) {
|
|
case PROP_DISPLAY:
|
|
g_object_get_property (G_OBJECT (mpeg_dec->src), "display", value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_dec_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstVdpMpegDec *mpeg_dec = (GstVdpMpegDec *) object;
|
|
|
|
switch (prop_id) {
|
|
case PROP_DISPLAY:
|
|
g_object_set_property (G_OBJECT (mpeg_dec->src), "display", value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_dec_base_init (gpointer gclass)
|
|
{
|
|
GstElementClass *element_class = GST_ELEMENT_CLASS (gclass);
|
|
|
|
GstCaps *src_caps;
|
|
GstPadTemplate *src_template;
|
|
|
|
gst_element_class_set_details_simple (element_class,
|
|
"VDPAU Mpeg Decoder",
|
|
"Decoder",
|
|
"decode mpeg stream with vdpau",
|
|
"Carl-Anton Ingmarsson <ca.ingmarsson@gmail.com>");
|
|
|
|
|
|
gst_element_class_add_pad_template (element_class,
|
|
gst_static_pad_template_get (&sink_template));
|
|
|
|
src_caps = gst_vdp_video_buffer_get_caps (TRUE, VDP_CHROMA_TYPE_420);
|
|
src_template = gst_pad_template_new ("src", GST_PAD_SRC, GST_PAD_ALWAYS,
|
|
src_caps);
|
|
|
|
gst_element_class_add_pad_template (element_class, src_template);
|
|
}
|
|
|
|
/* initialize the vdpaumpegdecoder's class */
|
|
static void
|
|
gst_vdp_mpeg_dec_class_init (GstVdpMpegDecClass * klass)
|
|
{
|
|
GObjectClass *gobject_class;
|
|
GstElementClass *gstelement_class;
|
|
|
|
gobject_class = (GObjectClass *) klass;
|
|
gstelement_class = (GstElementClass *) klass;
|
|
|
|
gobject_class->get_property = gst_vdp_mpeg_dec_get_property;
|
|
gobject_class->set_property = gst_vdp_mpeg_dec_set_property;
|
|
gobject_class->finalize = gst_vdp_mpeg_dec_finalize;
|
|
|
|
gstelement_class->change_state = gst_vdp_mpeg_dec_change_state;
|
|
|
|
g_object_class_install_property (gobject_class, PROP_DISPLAY,
|
|
g_param_spec_string ("display", "Display", "X Display name",
|
|
NULL, G_PARAM_READWRITE | G_PARAM_CONSTRUCT));
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_dec_init_info (VdpPictureInfoMPEG1Or2 * vdp_info)
|
|
{
|
|
vdp_info->forward_reference = VDP_INVALID_HANDLE;
|
|
vdp_info->backward_reference = VDP_INVALID_HANDLE;
|
|
vdp_info->slice_count = 0;
|
|
vdp_info->picture_structure = 3;
|
|
vdp_info->picture_coding_type = 0;
|
|
vdp_info->intra_dc_precision = 0;
|
|
vdp_info->frame_pred_frame_dct = 1;
|
|
vdp_info->concealment_motion_vectors = 0;
|
|
vdp_info->intra_vlc_format = 0;
|
|
vdp_info->alternate_scan = 0;
|
|
vdp_info->q_scale_type = 0;
|
|
vdp_info->top_field_first = 1;
|
|
}
|
|
|
|
static void
|
|
gst_vdp_mpeg_dec_init (GstVdpMpegDec * mpeg_dec, GstVdpMpegDecClass * gclass)
|
|
{
|
|
GstPadTemplate *src_template;
|
|
|
|
/* SRC PAD */
|
|
src_template =
|
|
gst_element_class_get_pad_template (GST_ELEMENT_CLASS (gclass), "src");
|
|
mpeg_dec->src =
|
|
GST_PAD (gst_vdp_video_src_pad_new (gst_pad_template_get_caps
|
|
(src_template)));
|
|
|
|
gst_pad_set_event_function (mpeg_dec->src,
|
|
GST_DEBUG_FUNCPTR (gst_vdp_mpeg_dec_src_event));
|
|
gst_pad_set_query_function (mpeg_dec->src,
|
|
GST_DEBUG_FUNCPTR (gst_vdp_mpeg_dec_src_query));
|
|
gst_pad_set_query_type_function (mpeg_dec->src,
|
|
GST_DEBUG_FUNCPTR (gst_mpeg_dec_get_querytypes));
|
|
gst_element_add_pad (GST_ELEMENT (mpeg_dec), mpeg_dec->src);
|
|
|
|
mpeg_dec->sink = gst_pad_new_from_static_template (&sink_template, "sink");
|
|
gst_pad_set_setcaps_function (mpeg_dec->sink,
|
|
GST_DEBUG_FUNCPTR (gst_vdp_mpeg_dec_set_caps));
|
|
gst_pad_set_chain_function (mpeg_dec->sink,
|
|
GST_DEBUG_FUNCPTR (gst_vdp_mpeg_dec_chain));
|
|
gst_pad_set_event_function (mpeg_dec->sink,
|
|
GST_DEBUG_FUNCPTR (gst_vdp_mpeg_dec_sink_event));
|
|
gst_element_add_pad (GST_ELEMENT (mpeg_dec), mpeg_dec->sink);
|
|
|
|
mpeg_dec->adapter = gst_adapter_new ();
|
|
mpeg_dec->mutex = g_mutex_new ();
|
|
}
|