vdpau: add new GstVdpauYUVVideo element

the GstVdpauYUVVideo element takes raw YUV video and outputs
GstVdpauVideoBuffers
This commit is contained in:
Carl-Anton Ingmarsson 2009-04-07 20:46:49 +02:00 committed by Jan Schmidt
parent 74485f6fa2
commit f8e0c0bcc1
7 changed files with 617 additions and 2 deletions

View file

@ -7,7 +7,8 @@ libgstvdpau_la_SOURCES = \
mpegutil.c \
gstvdpauvideoyuv.c \
gstvdpauvideobuffer.c \
gstvdpau.c
gstvdpau.c \
gstvdpauyuvvideo.c
libgstvdpau_la_CFLAGS = $(GST_CFLAGS) $(X11_CFLAGS) -Ivdpau
libgstvdpau_la_LIBADD = $(GST_LIBS) $(GST_BASE_LIBS) \
@ -19,8 +20,9 @@ noinst_HEADERS = \
gstvdpaudevice.h \
gstvdpaudecoder.h \
gstvdpaumpegdecoder.h \
mpegutil.h \
gstvdpauvideoyuv.h \
gstvdpauvideobuffer.h \
mpegutil.h
gstvdpauyuvvideo.h

View file

@ -7,6 +7,8 @@
#include "gstvdpaumpegdecoder.h"
#include "gstvdpauvideoyuv.h"
#include "gstvdpauyuvvideo.h"
static gboolean
vdpau_init (GstPlugin * vdpaumpegdecoder)
{
@ -14,6 +16,8 @@ vdpau_init (GstPlugin * vdpaumpegdecoder)
GST_RANK_NONE, GST_TYPE_VDPAU_MPEG_DECODER);
gst_element_register (vdpaumpegdecoder, "vdpauvideoyuv",
GST_RANK_NONE, GST_TYPE_VDPAU_VIDEO_YUV);
gst_element_register (vdpaumpegdecoder, "vdpauyuvvideo",
GST_RANK_NONE, GST_TYPE_VDPAU_YUV_VIDEO);
return TRUE;
}

View file

@ -83,6 +83,8 @@ gst_vdpau_device_constructed (GObject * object)
&device->vdp_video_surface_query_ycbcr_capabilities},
{VDP_FUNC_ID_VIDEO_SURFACE_GET_BITS_Y_CB_CR,
&device->vdp_video_surface_get_bits_ycbcr},
{VDP_FUNC_ID_VIDEO_SURFACE_PUT_BITS_Y_CB_CR,
&device->vdp_video_surface_put_bits_ycbcr},
{VDP_FUNC_ID_VIDEO_SURFACE_GET_PARAMETERS,
&device->vdp_video_surface_get_parameters},
{VDP_FUNC_ID_DECODER_CREATE, &device->vdp_decoder_create},
@ -257,6 +259,10 @@ gst_vdpau_get_device (const gchar * display_name)
if (!device) {
device = gst_vdpau_device_new (display_name);
g_object_weak_ref (G_OBJECT (device), device_destroyed_cb, devices_hash);
if (display_name)
g_hash_table_insert (devices_hash, g_strdup (display_name), device);
else
g_hash_table_insert (devices_hash, g_strdup (""), device);
} else
g_object_ref (device);

View file

@ -61,6 +61,7 @@ struct _GstVdpauDevice
VdpVideoSurfaceQueryGetPutBitsYCbCrCapabilities *vdp_video_surface_query_ycbcr_capabilities;
VdpVideoSurfaceGetParameters *vdp_video_surface_get_parameters;
VdpVideoSurfaceGetBitsYCbCr *vdp_video_surface_get_bits_ycbcr;
VdpVideoSurfacePutBitsYCbCr *vdp_video_surface_put_bits_ycbcr;
VdpDecoderCreate *vdp_decoder_create;
VdpDecoderDestroy *vdp_decoder_destroy;

View file

@ -412,6 +412,7 @@ gst_vdpau_video_yuv_init (GstVdpauVideoYUV * video_yuv,
gst_vdpau_video_yuv_sink_set_caps);
gst_pad_set_chain_function (video_yuv->sink, gst_vdpau_video_yuv_chain);
gst_element_add_pad (GST_ELEMENT (video_yuv), video_yuv->sink);
gst_pad_set_active (video_yuv->sink, TRUE);
}
static void

View file

@ -0,0 +1,537 @@
/*
* GStreamer
* Copyright (C) 2009 Carl-Anton Ingmarsson <ca.ingmarsson@gmail.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
* Boston, MA 02111-1307, USA.
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#endif
#include <gst/gst.h>
#include <gst/video/video.h>
#include "gstvdpauvideobuffer.h"
#include "gstvdpauyuvvideo.h"
GST_DEBUG_CATEGORY_STATIC (gst_vdpau_yuv_video_debug);
#define GST_CAT_DEFAULT gst_vdpau_yuv_video_debug
/* Filter signals and args */
enum
{
/* FILL ME */
LAST_SIGNAL
};
enum
{
PROP_0,
PROP_DISPLAY,
PROP_SILENT
};
static GstStaticPadTemplate sink_template = GST_STATIC_PAD_TEMPLATE ("sink",
GST_PAD_SINK,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("video/x-raw-yuv, "
"framerate = (fraction) [ 0, MAX ], "
"width = (int) [ 1, MAX ], " "height = (int) [ 1, MAX ]"));
static GstStaticPadTemplate src_template = GST_STATIC_PAD_TEMPLATE ("src",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS (GST_VDPAU_VIDEO_CAPS));
#define DEBUG_INIT(bla) \
GST_DEBUG_CATEGORY_INIT (gst_vdpau_yuv_video_debug, "vdpauvideoyuv", 0, "YUV to VDPAU video surface");
GST_BOILERPLATE_FULL (GstVdpauYUVVideo, gst_vdpau_yuv_video, GstElement,
GST_TYPE_ELEMENT, DEBUG_INIT);
static void gst_vdpau_yuv_video_finalize (GObject * object);
static void gst_vdpau_yuv_video_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec);
static void gst_vdpau_yuv_video_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec);
GstFlowReturn
gst_vdpau_yuv_video_chain (GstPad * pad, GstBuffer * buffer)
{
GstVdpauYUVVideo *yuv_video;
GstVdpauDevice *device;
VdpVideoSurface surface;
GstBuffer *outbuf = NULL;
yuv_video = GST_VDPAU_YUV_VIDEO (GST_OBJECT_PARENT (pad));
device = yuv_video->device;
outbuf =
GST_BUFFER (gst_vdpau_video_buffer_new (device, yuv_video->chroma_type,
yuv_video->width, yuv_video->height));
surface = GST_VDPAU_VIDEO_BUFFER (outbuf)->surface;
switch (yuv_video->format) {
case GST_MAKE_FOURCC ('Y', 'V', '1', '2'):
{
VdpStatus status;
guint8 *data[3];
guint32 stride[3];
data[0] = GST_BUFFER_DATA (buffer) +
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_YV12,
0, yuv_video->width, yuv_video->height);
data[1] = data[0] +
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_YV12,
2, yuv_video->width, yuv_video->height);
data[2] = data[0] +
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_YV12,
1, yuv_video->width, yuv_video->height);
stride[0] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_YV12,
0, yuv_video->width);
stride[1] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_YV12,
2, yuv_video->width);
stride[2] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_YV12,
1, yuv_video->width);
status =
device->vdp_video_surface_put_bits_ycbcr (surface,
VDP_YCBCR_FORMAT_YV12, (void *) data, stride);
if (G_UNLIKELY (status != VDP_STATUS_OK)) {
GST_ELEMENT_ERROR (yuv_video, RESOURCE, READ,
("Couldn't push YV12 data to VDPAU"),
("Error returned from vdpau was: %s",
device->vdp_get_error_string (status)));
break;
}
break;
}
case GST_MAKE_FOURCC ('I', '4', '2', '0'):
{
VdpStatus status;
guint8 *data[3];
guint32 stride[3];
data[0] = GST_BUFFER_DATA (buffer) +
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420,
0, yuv_video->width, yuv_video->height);
data[1] = GST_BUFFER_DATA (buffer) +
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420,
2, yuv_video->width, yuv_video->height);
data[2] = GST_BUFFER_DATA (buffer) +
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420,
1, yuv_video->width, yuv_video->height);
stride[0] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420,
0, yuv_video->width);
stride[1] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420,
2, yuv_video->width);
stride[2] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420,
1, yuv_video->width);
status =
device->vdp_video_surface_put_bits_ycbcr (surface,
VDP_YCBCR_FORMAT_YV12, (void *) data, stride);
if (G_UNLIKELY (status != VDP_STATUS_OK)) {
GST_ELEMENT_ERROR (yuv_video, RESOURCE, READ,
("Couldn't push YV12 data to VDPAU"),
("Error returned from vdpau was: %s",
device->vdp_get_error_string (status)));
break;
}
break;
}
case GST_MAKE_FOURCC ('N', 'V', '1', '2'):
{
VdpStatus status;
guint8 *data[2];
guint32 stride[2];
data[0] = GST_BUFFER_DATA (buffer);
data[1] = GST_BUFFER_DATA (buffer) + yuv_video->width * yuv_video->height;
stride[0] = yuv_video->width;
stride[1] = yuv_video->width;
status =
device->vdp_video_surface_put_bits_ycbcr (surface,
VDP_YCBCR_FORMAT_NV12, (void *) data, stride);
if (G_UNLIKELY (status != VDP_STATUS_OK)) {
GST_ELEMENT_ERROR (yuv_video, RESOURCE, READ,
("Couldn't get data from vdpau"),
("Error returned from vdpau was: %s",
device->vdp_get_error_string (status)));
break;
}
break;
}
default:
break;
}
gst_buffer_unref (buffer);
if (outbuf) {
gst_buffer_copy_metadata (outbuf, buffer, GST_BUFFER_COPY_TIMESTAMPS);
gst_buffer_set_caps (outbuf, GST_PAD_CAPS (yuv_video->src));
return gst_pad_push (yuv_video->src, outbuf);
}
return GST_FLOW_ERROR;
}
typedef struct
{
VdpChromaType chroma_type;
VdpYCbCrFormat format;
guint32 fourcc;
} VdpauFormats;
static VdpChromaType chroma_types[3] =
{ VDP_CHROMA_TYPE_420, VDP_CHROMA_TYPE_422, VDP_CHROMA_TYPE_444 };
static VdpauFormats formats[7] = {
{
VDP_CHROMA_TYPE_420,
VDP_YCBCR_FORMAT_NV12,
GST_MAKE_FOURCC ('N', 'V', '1', '2')
},
{
VDP_CHROMA_TYPE_422,
VDP_YCBCR_FORMAT_UYVY,
GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y')
},
{
VDP_CHROMA_TYPE_444,
VDP_YCBCR_FORMAT_V8U8Y8A8,
GST_MAKE_FOURCC ('A', 'Y', 'U', 'V')
},
{
VDP_CHROMA_TYPE_444,
VDP_YCBCR_FORMAT_Y8U8V8A8,
GST_MAKE_FOURCC ('A', 'V', 'U', 'Y')
},
{
VDP_CHROMA_TYPE_422,
VDP_YCBCR_FORMAT_YUYV,
GST_MAKE_FOURCC ('Y', 'U', 'Y', 'V')
},
{
VDP_CHROMA_TYPE_420,
VDP_YCBCR_FORMAT_YV12,
GST_MAKE_FOURCC ('Y', 'V', '1', '2')
},
{
VDP_CHROMA_TYPE_420,
VDP_YCBCR_FORMAT_YV12,
GST_MAKE_FOURCC ('I', '4', '2', '0')
}
};
static GstCaps *
gst_vdpau_yuv_video_get_caps (GstVdpauYUVVideo * yuv_video)
{
GstVdpauDevice *device;
GstCaps *caps;
gint i;
device = yuv_video->device;
caps = gst_caps_new_empty ();
for (i = 0; i < 3; i++) {
VdpStatus status;
VdpBool is_supported;
guint32 max_w, max_h;
status =
device->vdp_video_surface_query_capabilities (device->device,
chroma_types[i], &is_supported, &max_w, &max_h);
if (status != VDP_STATUS_OK && status != VDP_STATUS_INVALID_CHROMA_TYPE) {
GST_ELEMENT_ERROR (yuv_video, RESOURCE, READ,
("Could not get query VDPAU video surface capabilites"),
("Error returned from vdpau was: %s",
device->vdp_get_error_string (status)));
goto error;
}
if (is_supported) {
gint j;
for (j = 0; j < 7; j++) {
if (formats[j].chroma_type != chroma_types[i])
continue;
status =
device->vdp_video_surface_query_ycbcr_capabilities (device->device,
formats[j].chroma_type, formats[j].format, &is_supported);
if (status != VDP_STATUS_OK
&& status != VDP_STATUS_INVALID_Y_CB_CR_FORMAT) {
GST_ELEMENT_ERROR (yuv_video, RESOURCE, READ,
("Could not query VDPAU YCbCr capabilites"),
("Error returned from vdpau was: %s",
device->vdp_get_error_string (status)));
goto error;
}
if (is_supported) {
GstCaps *format_caps;
format_caps = gst_caps_new_simple ("video/x-raw-yuv",
"format", GST_TYPE_FOURCC, formats[j].fourcc,
"width", GST_TYPE_INT_RANGE, 1, max_w,
"height", GST_TYPE_INT_RANGE, 1, max_h,
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, G_MAXINT, 1, NULL);
gst_caps_append (caps, format_caps);
}
}
}
}
error:
if (gst_caps_is_empty (caps)) {
gst_caps_unref (caps);
return NULL;
}
return caps;
}
static gboolean
gst_vdpau_yuv_video_sink_setcaps (GstPad * pad, GstCaps * caps)
{
GstVdpauYUVVideo *yuv_video = GST_VDPAU_YUV_VIDEO (GST_OBJECT_PARENT (pad));
GstStructure *structure;
guint32 fourcc;
gint chroma_type = 0;
gint width, height;
gint framerate_numerator, framerate_denominator;
gint par_numerator, par_denominator;
gint i;
GstCaps *src_caps, *new_caps;
gboolean res;
structure = gst_caps_get_structure (caps, 0);
gst_structure_get_fourcc (structure, "format", &fourcc);
gst_structure_get_int (structure, "width", &width);
gst_structure_get_int (structure, "height", &height);
gst_structure_get_fraction (structure, "framerate",
&framerate_numerator, &framerate_denominator);
gst_structure_get_fraction (structure, "pixel-aspect-ratio",
&par_numerator, &par_denominator);
for (i = 0; i < 7; i++) {
if (formats[i].fourcc == fourcc) {
chroma_type = formats[i].chroma_type;
break;
}
}
src_caps = gst_pad_get_allowed_caps (yuv_video->src);
if (G_UNLIKELY (!src_caps || !gst_caps_get_size (src_caps)))
return FALSE;
new_caps = gst_caps_copy_nth (src_caps, 0);
gst_caps_unref (src_caps);
if (G_UNLIKELY (!new_caps))
return FALSE;
structure = gst_caps_get_structure (new_caps, 0);
gst_structure_set (structure,
"device", G_TYPE_OBJECT, yuv_video->device,
"chroma-type", G_TYPE_INT, chroma_type,
"width", G_TYPE_INT, width,
"height", G_TYPE_INT, height,
"framerate", GST_TYPE_FRACTION, framerate_numerator,
framerate_denominator, "pixel-aspect-ratio", GST_TYPE_FRACTION,
par_numerator, par_denominator, NULL);
gst_pad_fixate_caps (yuv_video->src, new_caps);
res = gst_pad_set_caps (yuv_video->src, new_caps);
gst_caps_unref (new_caps);
if (G_UNLIKELY (!res))
return FALSE;
yuv_video->width = width;
yuv_video->height = height;
yuv_video->format = fourcc;
yuv_video->chroma_type = chroma_type;
return TRUE;
}
static GstCaps *
gst_vdpau_yuv_video_sink_getcaps (GstPad * pad)
{
GstVdpauYUVVideo *yuv_video;
yuv_video = GST_VDPAU_YUV_VIDEO (GST_OBJECT_PARENT (pad));
if (yuv_video->sink_caps)
return gst_caps_copy (yuv_video->sink_caps);
return gst_caps_copy (gst_pad_get_pad_template_caps (yuv_video->sink));
}
static GstStateChangeReturn
gst_vdpau_yuv_video_change_state (GstElement * element,
GstStateChange transition)
{
GstVdpauYUVVideo *yuv_video;
yuv_video = GST_VDPAU_YUV_VIDEO (element);
switch (transition) {
case GST_STATE_CHANGE_NULL_TO_READY:
yuv_video->device = gst_vdpau_get_device (yuv_video->display);
if (!yuv_video->sink_caps)
yuv_video->sink_caps = gst_vdpau_yuv_video_get_caps (yuv_video);
break;
case GST_STATE_CHANGE_READY_TO_NULL:
g_object_unref (yuv_video->device);
yuv_video->device = NULL;
break;
default:
break;
}
return GST_STATE_CHANGE_SUCCESS;
}
/* GObject vmethod implementations */
static void
gst_vdpau_yuv_video_base_init (gpointer klass)
{
GstElementClass *element_class = GST_ELEMENT_CLASS (klass);
gst_element_class_set_details_simple (element_class,
"VdpauYUVVideo",
"Coyuv_video/Decoder/Video",
"VDPAU video surface to YUV",
"Carl-Anton Ingmarsson <ca.ingmarsson@gmail.com>");
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&sink_template));
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&src_template));
}
static void
gst_vdpau_yuv_video_class_init (GstVdpauYUVVideoClass * klass)
{
GObjectClass *gobject_class;
GstElementClass *gstelement_class;
gobject_class = (GObjectClass *) klass;
gstelement_class = (GstElementClass *) klass;
gobject_class->finalize = gst_vdpau_yuv_video_finalize;
gobject_class->set_property = gst_vdpau_yuv_video_set_property;
gobject_class->get_property = gst_vdpau_yuv_video_get_property;
g_object_class_install_property (gobject_class, PROP_DISPLAY,
g_param_spec_string ("display", "Display", "X Display name",
NULL, G_PARAM_READWRITE | G_PARAM_CONSTRUCT));
g_object_class_install_property (gobject_class, PROP_SILENT,
g_param_spec_boolean ("silent", "Silent", "Produce verbose output ?",
FALSE, G_PARAM_READWRITE | GST_PARAM_CONTROLLABLE));
gstelement_class->change_state = gst_vdpau_yuv_video_change_state;
}
static void
gst_vdpau_yuv_video_init (GstVdpauYUVVideo * yuv_video,
GstVdpauYUVVideoClass * klass)
{
yuv_video->silent = FALSE;
yuv_video->sink_caps = NULL;
yuv_video->display = NULL;
yuv_video->device = NULL;
yuv_video->height = 0;
yuv_video->width = 0;
yuv_video->format = 0;
yuv_video->chroma_type = 0;
yuv_video->src = gst_pad_new_from_static_template (&src_template, "src");
gst_element_add_pad (GST_ELEMENT (yuv_video), yuv_video->src);
yuv_video->sink = gst_pad_new_from_static_template (&sink_template, "sink");
gst_pad_set_getcaps_function (yuv_video->sink,
gst_vdpau_yuv_video_sink_getcaps);
gst_pad_set_setcaps_function (yuv_video->sink,
gst_vdpau_yuv_video_sink_setcaps);
gst_pad_set_chain_function (yuv_video->sink, gst_vdpau_yuv_video_chain);
gst_element_add_pad (GST_ELEMENT (yuv_video), yuv_video->sink);
gst_pad_set_active (yuv_video->sink, TRUE);
}
static void
gst_vdpau_yuv_video_finalize (GObject * object)
{
GstVdpauYUVVideo *yuv_video = (GstVdpauYUVVideo *) object;
g_free (yuv_video->display);
}
static void
gst_vdpau_yuv_video_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec)
{
GstVdpauYUVVideo *yuv_video = GST_VDPAU_YUV_VIDEO (object);
switch (prop_id) {
case PROP_DISPLAY:
g_free (yuv_video->display);
yuv_video->display = g_value_dup_string (value);
break;
case PROP_SILENT:
yuv_video->silent = g_value_get_boolean (value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static void
gst_vdpau_yuv_video_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec)
{
GstVdpauYUVVideo *yuv_video = GST_VDPAU_YUV_VIDEO (object);
switch (prop_id) {
case PROP_DISPLAY:
g_value_set_string (value, yuv_video->display);
break;
case PROP_SILENT:
g_value_set_boolean (value, yuv_video->silent);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}

View file

@ -0,0 +1,64 @@
/*
* GStreamer
* Copyright (C) 2009 Carl-Anton Ingmarsson <ca.ingmarsson@gmail.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
* Boston, MA 02111-1307, USA.
*/
#ifndef __GST_VDPAU_YUV_VIDEO_H__
#define __GST_VDPAU_YUV_VIDEO_H__
#include <gst/gst.h>
#include "gstvdpaudevice.h"
G_BEGIN_DECLS
#define GST_TYPE_VDPAU_YUV_VIDEO (gst_vdpau_yuv_video_get_type())
#define GST_VDPAU_YUV_VIDEO(obj) (G_TYPE_CHECK_INSTANCE_CAST((obj),GST_TYPE_VDPAU_YUV_VIDEO,GstVdpauYUVVideo))
#define GST_VDPAU_YUV_VIDEO_CLASS(klass) (G_TYPE_CHECK_CLASS_CAST((klass),GST_TYPE_VDPAU_YUV_VIDEO,GstVdpauYUVVideoClass))
#define GST_VDPAU_YUV_VIDEO_GET_CLASS(obj) (G_TYPE_INSTANCE_GET_CLASS ((obj), GST_TYPE_VDPAU_YUV_VIDEO, GstVdpauYUVVideoClass))
#define GST_IS_VDPAU_YUV_VIDEO(obj) (G_TYPE_CHECK_INSTANCE_TYPE((obj),GST_TYPE_VDPAU_YUV_VIDEO))
#define GST_IS_VDPAU_YUV_VIDEO_CLASS(klass) (G_TYPE_CHECK_CLASS_TYPE((klass),GST_TYPE_VDPAU_YUV_VIDEO))
typedef struct _GstVdpauYUVVideo GstVdpauYUVVideo;
typedef struct _GstVdpauYUVVideoClass GstVdpauYUVVideoClass;
struct _GstVdpauYUVVideo {
GstElement element;
GstPad *src, *sink;
GstCaps *sink_caps;
gchar *display;
GstVdpauDevice *device;
guint32 format;
gint chroma_type;
gint width, height;
gboolean silent;
};
struct _GstVdpauYUVVideoClass {
GstElementClass parent_class;
};
GType gst_vdpau_yuv_video_get_type (void);
G_END_DECLS
#endif /* __GST_VDPAU_YUV_VIDEO_H__ */