videofilter: improve video filter

Flesh out the video filter base class. Make it parse the input and output caps
and turn them into GstVideoInfo. Map buffers as video frames and pass them to
the transform functions.
This allows us to also implement the propose and decide_allocation vmethods.
Implement the transform size method as well.
Update subclasses with the new improvements.
This commit is contained in:
Wim Taymans 2011-12-21 23:46:53 +01:00
parent 8a9c4ed69f
commit a5f3d21723
6 changed files with 331 additions and 393 deletions

View file

@ -40,37 +40,98 @@
#include "gstvideofilter.h" #include "gstvideofilter.h"
#include <gst/video/video.h> #include <gst/video/video.h>
#include <gst/video/gstvideometa.h>
#include <gst/video/gstvideopool.h>
GST_DEBUG_CATEGORY_STATIC (gst_video_filter_debug); GST_DEBUG_CATEGORY_STATIC (gst_video_filter_debug);
#define GST_CAT_DEFAULT gst_video_filter_debug #define GST_CAT_DEFAULT gst_video_filter_debug
static void gst_video_filter_class_init (gpointer g_class, gpointer class_data); #define gst_video_filter_parent_class parent_class
static void gst_video_filter_init (GTypeInstance * instance, gpointer g_class); G_DEFINE_ABSTRACT_TYPE (GstVideoFilter, gst_video_filter,
GST_TYPE_BASE_TRANSFORM);
static GstBaseTransformClass *parent_class = NULL; /* Answer the allocation query downstream. This is only called for
* non-passthrough cases */
GType static gboolean
gst_video_filter_get_type (void) gst_video_filter_propose_allocation (GstBaseTransform * trans, GstQuery * query)
{ {
static GType video_filter_type = 0; GstVideoFilter *filter = GST_VIDEO_FILTER_CAST (trans);
GstBufferPool *pool;
GstCaps *caps;
gboolean need_pool;
guint size;
if (!video_filter_type) { gst_query_parse_allocation (query, &caps, &need_pool);
static const GTypeInfo video_filter_info = {
sizeof (GstVideoFilterClass),
NULL,
NULL,
gst_video_filter_class_init,
NULL,
NULL,
sizeof (GstVideoFilter),
0,
gst_video_filter_init,
};
video_filter_type = g_type_register_static (GST_TYPE_BASE_TRANSFORM, size = GST_VIDEO_INFO_SIZE (&filter->in_info);
"GstVideoFilter", &video_filter_info, G_TYPE_FLAG_ABSTRACT);
if (need_pool) {
GstStructure *structure;
pool = gst_video_buffer_pool_new ();
structure = gst_buffer_pool_get_config (pool);
gst_buffer_pool_config_set (structure, caps, size, 0, 0, 0, 15);
if (!gst_buffer_pool_set_config (pool, structure))
goto config_failed;
} else
pool = NULL;
gst_query_set_allocation_params (query, size, 0, 0, 0, 15, pool);
gst_object_unref (pool);
gst_query_add_allocation_meta (query, GST_VIDEO_META_API);
return TRUE;
/* ERRORS */
config_failed:
{
GST_ERROR_OBJECT (filter, "failed to set config");
gst_object_unref (pool);
return FALSE;
} }
return video_filter_type; }
/* configure the allocation query that was answered downstream, we can configure
* some properties on it. Only called in passthrough mode. */
static gboolean
gst_video_filter_decide_allocation (GstBaseTransform * trans, GstQuery * query)
{
GstBufferPool *pool = NULL;
guint size, min, max, prefix, alignment;
gst_query_parse_allocation_params (query, &size, &min, &max, &prefix,
&alignment, &pool);
if (pool) {
GstStructure *config;
config = gst_buffer_pool_get_config (pool);
gst_buffer_pool_config_add_option (config,
GST_BUFFER_POOL_OPTION_VIDEO_META);
gst_buffer_pool_set_config (pool, config);
}
return TRUE;
}
/* our output size only depends on the caps, not on the input caps */
static gboolean
gst_video_filter_transform_size (GstBaseTransform * btrans,
GstPadDirection direction, GstCaps * caps, gsize size,
GstCaps * othercaps, gsize * othersize)
{
gboolean ret = TRUE;
GstVideoInfo info;
g_assert (size);
ret = gst_video_info_from_caps (&info, othercaps);
if (ret)
*othersize = info.size;
return ret;
} }
static gboolean static gboolean
@ -92,8 +153,134 @@ gst_video_filter_get_unit_size (GstBaseTransform * btrans, GstCaps * caps,
return TRUE; return TRUE;
} }
static gboolean
gst_video_filter_set_caps (GstBaseTransform * trans, GstCaps * incaps,
GstCaps * outcaps)
{
GstVideoFilter *filter = GST_VIDEO_FILTER_CAST (trans);
GstVideoFilterClass *fclass;
GstVideoInfo in_info, out_info;
gboolean res;
/* input caps */
if (!gst_video_info_from_caps (&in_info, incaps))
goto invalid_caps;
/* output caps */
if (!gst_video_info_from_caps (&out_info, outcaps))
goto invalid_caps;
fclass = GST_VIDEO_FILTER_GET_CLASS (filter);
if (fclass->set_info)
res = fclass->set_info (filter, incaps, &in_info, outcaps, &out_info);
else
res = TRUE;
if (res) {
filter->in_info = in_info;
filter->out_info = out_info;
}
filter->negotiated = res;
return res;
/* ERRORS */
invalid_caps:
{
GST_ERROR_OBJECT (filter, "invalid caps");
filter->negotiated = FALSE;
return FALSE;
}
}
static GstFlowReturn
gst_video_filter_transform (GstBaseTransform * trans, GstBuffer * inbuf,
GstBuffer * outbuf)
{
GstFlowReturn res;
GstVideoFilter *filter = GST_VIDEO_FILTER_CAST (trans);
GstVideoFilterClass *fclass;
if (G_UNLIKELY (!filter->negotiated))
goto unknown_format;
fclass = GST_VIDEO_FILTER_GET_CLASS (filter);
if (fclass->transform_frame) {
GstVideoFrame in_frame, out_frame;
if (!gst_video_frame_map (&in_frame, &filter->in_info, inbuf, GST_MAP_READ))
goto invalid_buffer;
if (!gst_video_frame_map (&out_frame, &filter->out_info, outbuf,
GST_MAP_WRITE))
goto invalid_buffer;
res = fclass->transform_frame (filter, &in_frame, &out_frame);
gst_video_frame_unmap (&out_frame);
gst_video_frame_unmap (&in_frame);
} else
res = GST_FLOW_OK;
return res;
/* ERRORS */
unknown_format:
{
GST_ELEMENT_ERROR (filter, CORE, NOT_IMPLEMENTED, (NULL),
("unknown format"));
return GST_FLOW_NOT_NEGOTIATED;
}
invalid_buffer:
{
GST_ELEMENT_WARNING (filter, CORE, NOT_IMPLEMENTED, (NULL),
("invalid video buffer received"));
return GST_FLOW_OK;
}
}
static GstFlowReturn
gst_video_filter_transform_ip (GstBaseTransform * trans, GstBuffer * buf)
{
GstFlowReturn res;
GstVideoFilter *filter = GST_VIDEO_FILTER_CAST (trans);
GstVideoFilterClass *fclass;
if (G_UNLIKELY (!filter->negotiated))
goto unknown_format;
fclass = GST_VIDEO_FILTER_GET_CLASS (filter);
if (fclass->transform_frame_ip) {
GstVideoFrame frame;
if (!gst_video_frame_map (&frame, &filter->in_info, buf, GST_MAP_READWRITE))
goto invalid_buffer;
res = fclass->transform_frame_ip (filter, &frame);
gst_video_frame_unmap (&frame);
} else
res = GST_FLOW_OK;
return res;
/* ERRORS */
unknown_format:
{
GST_ELEMENT_ERROR (filter, CORE, NOT_IMPLEMENTED, (NULL),
("unknown format"));
return GST_FLOW_NOT_NEGOTIATED;
}
invalid_buffer:
{
GST_ELEMENT_WARNING (filter, CORE, NOT_IMPLEMENTED, (NULL),
("invalid video buffer received"));
return GST_FLOW_OK;
}
}
static void static void
gst_video_filter_class_init (gpointer g_class, gpointer class_data) gst_video_filter_class_init (GstVideoFilterClass * g_class)
{ {
GstBaseTransformClass *trans_class; GstBaseTransformClass *trans_class;
GstVideoFilterClass *klass; GstVideoFilterClass *klass;
@ -101,23 +288,30 @@ gst_video_filter_class_init (gpointer g_class, gpointer class_data)
klass = (GstVideoFilterClass *) g_class; klass = (GstVideoFilterClass *) g_class;
trans_class = (GstBaseTransformClass *) klass; trans_class = (GstBaseTransformClass *) klass;
trans_class->set_caps = GST_DEBUG_FUNCPTR (gst_video_filter_set_caps);
trans_class->propose_allocation =
GST_DEBUG_FUNCPTR (gst_video_filter_propose_allocation);
trans_class->decide_allocation =
GST_DEBUG_FUNCPTR (gst_video_filter_decide_allocation);
trans_class->transform_size =
GST_DEBUG_FUNCPTR (gst_video_filter_transform_size);
trans_class->get_unit_size = trans_class->get_unit_size =
GST_DEBUG_FUNCPTR (gst_video_filter_get_unit_size); GST_DEBUG_FUNCPTR (gst_video_filter_get_unit_size);
trans_class->transform = GST_DEBUG_FUNCPTR (gst_video_filter_transform);
parent_class = g_type_class_peek_parent (klass); trans_class->transform_ip = GST_DEBUG_FUNCPTR (gst_video_filter_transform_ip);
GST_DEBUG_CATEGORY_INIT (gst_video_filter_debug, "videofilter", 0, GST_DEBUG_CATEGORY_INIT (gst_video_filter_debug, "videofilter", 0,
"videofilter"); "videofilter");
} }
static void static void
gst_video_filter_init (GTypeInstance * instance, gpointer g_class) gst_video_filter_init (GstVideoFilter * instance)
{ {
GstVideoFilter *videofilter = GST_VIDEO_FILTER (instance); GstVideoFilter *videofilter = GST_VIDEO_FILTER (instance);
GST_DEBUG_OBJECT (videofilter, "gst_video_filter_init"); GST_DEBUG_OBJECT (videofilter, "gst_video_filter_init");
videofilter->inited = FALSE; videofilter->negotiated = FALSE;
/* enable QoS */ /* enable QoS */
gst_base_transform_set_qos_enabled (GST_BASE_TRANSFORM (videofilter), TRUE); gst_base_transform_set_qos_enabled (GST_BASE_TRANSFORM (videofilter), TRUE);
} }

View file

@ -22,6 +22,7 @@
#define __GST_VIDEO_FILTER_H__ #define __GST_VIDEO_FILTER_H__
#include <gst/base/gstbasetransform.h> #include <gst/base/gstbasetransform.h>
#include <gst/video/video.h>
G_BEGIN_DECLS G_BEGIN_DECLS
@ -40,11 +41,14 @@ typedef struct _GstVideoFilterClass GstVideoFilterClass;
(G_TYPE_CHECK_INSTANCE_TYPE((obj),GST_TYPE_VIDEO_FILTER)) (G_TYPE_CHECK_INSTANCE_TYPE((obj),GST_TYPE_VIDEO_FILTER))
#define GST_IS_VIDEO_FILTER_CLASS(klass) \ #define GST_IS_VIDEO_FILTER_CLASS(klass) \
(G_TYPE_CHECK_CLASS_TYPE((klass),GST_TYPE_VIDEO_FILTER)) (G_TYPE_CHECK_CLASS_TYPE((klass),GST_TYPE_VIDEO_FILTER))
#define GST_VIDEO_FILTER_CAST(obj) ((GstVideoFilter *)(obj))
struct _GstVideoFilter { struct _GstVideoFilter {
GstBaseTransform element; GstBaseTransform element;
gboolean inited; gboolean negotiated;
GstVideoInfo in_info;
GstVideoInfo out_info;
}; };
/** /**
@ -55,6 +59,15 @@ struct _GstVideoFilter {
*/ */
struct _GstVideoFilterClass { struct _GstVideoFilterClass {
GstBaseTransformClass parent_class; GstBaseTransformClass parent_class;
gboolean (*set_info) (GstVideoFilter *filter,
GstCaps *incaps, GstVideoInfo *in_info,
GstCaps *outcaps, GstVideoInfo *out_info);
/* transform */
GstFlowReturn (*transform_frame) (GstVideoFilter *filter,
GstVideoFrame *inframe, GstVideoFrame *outframe);
GstFlowReturn (*transform_frame_ip) (GstVideoFilter *trans, GstVideoFrame *frame);
}; };
GType gst_video_filter_get_type (void); GType gst_video_filter_get_type (void);

View file

@ -38,6 +38,7 @@
#endif #endif
#include "gstvideoconvert.h" #include "gstvideoconvert.h"
#include <gst/video/video.h> #include <gst/video/video.h>
#include <gst/video/gstvideometa.h> #include <gst/video/gstvideometa.h>
#include <gst/video/gstvideopool.h> #include <gst/video/gstvideopool.h>
@ -77,13 +78,11 @@ static void gst_video_convert_set_property (GObject * object,
static void gst_video_convert_get_property (GObject * object, static void gst_video_convert_get_property (GObject * object,
guint property_id, GValue * value, GParamSpec * pspec); guint property_id, GValue * value, GParamSpec * pspec);
static gboolean gst_video_convert_set_caps (GstBaseTransform * btrans, static gboolean gst_video_convert_set_info (GstVideoFilter * filter,
GstCaps * incaps, GstCaps * outcaps); GstCaps * incaps, GstVideoInfo * in_info, GstCaps * outcaps,
static gboolean gst_video_convert_transform_size (GstBaseTransform * btrans, GstVideoInfo * out_info);
GstPadDirection direction, GstCaps * caps, gsize size, static GstFlowReturn gst_video_convert_transform_frame (GstVideoFilter * filter,
GstCaps * othercaps, gsize * othersize); GstVideoFrame * in_frame, GstVideoFrame * out_frame);
static GstFlowReturn gst_video_convert_transform (GstBaseTransform * btrans,
GstBuffer * inbuf, GstBuffer * outbuf);
static GType static GType
dither_method_get_type (void) dither_method_get_type (void)
@ -159,96 +158,26 @@ gst_video_convert_transform_caps (GstBaseTransform * btrans,
return result; return result;
} }
/* Answer the allocation query downstream. This is only called for
* non-passthrough cases */
static gboolean static gboolean
gst_video_convert_propose_allocation (GstBaseTransform * trans, gst_video_convert_set_info (GstVideoFilter * filter,
GstQuery * query) GstCaps * incaps, GstVideoInfo * in_info, GstCaps * outcaps,
{ GstVideoInfo * out_info)
GstVideoConvert *space = GST_VIDEO_CONVERT_CAST (trans);
GstBufferPool *pool;
GstCaps *caps;
gboolean need_pool;
guint size;
gst_query_parse_allocation (query, &caps, &need_pool);
size = GST_VIDEO_INFO_SIZE (&space->from_info);
if (need_pool) {
GstStructure *structure;
pool = gst_video_buffer_pool_new ();
structure = gst_buffer_pool_get_config (pool);
gst_buffer_pool_config_set (structure, caps, size, 0, 0, 0, 15);
if (!gst_buffer_pool_set_config (pool, structure))
goto config_failed;
} else
pool = NULL;
gst_query_set_allocation_params (query, size, 0, 0, 0, 15, pool);
gst_object_unref (pool);
gst_query_add_allocation_meta (query, GST_VIDEO_META_API);
return TRUE;
/* ERRORS */
config_failed:
{
GST_ERROR_OBJECT (space, "failed to set config.");
gst_object_unref (pool);
return FALSE;
}
}
/* configure the allocation query that was answered downstream, we can configure
* some properties on it. Only called in passthrough mode. */
static gboolean
gst_video_convert_decide_allocation (GstBaseTransform * trans, GstQuery * query)
{
GstBufferPool *pool = NULL;
guint size, min, max, prefix, alignment;
gst_query_parse_allocation_params (query, &size, &min, &max, &prefix,
&alignment, &pool);
if (pool) {
GstStructure *config;
config = gst_buffer_pool_get_config (pool);
gst_buffer_pool_config_add_option (config,
GST_BUFFER_POOL_OPTION_VIDEO_META);
gst_buffer_pool_set_config (pool, config);
}
return TRUE;
}
static gboolean
gst_video_convert_set_caps (GstBaseTransform * btrans, GstCaps * incaps,
GstCaps * outcaps)
{ {
GstVideoConvert *space; GstVideoConvert *space;
GstVideoInfo in_info;
GstVideoInfo out_info;
ColorSpaceColorSpec in_spec, out_spec; ColorSpaceColorSpec in_spec, out_spec;
gboolean interlaced; gboolean interlaced;
space = GST_VIDEO_CONVERT_CAST (btrans); space = GST_VIDEO_CONVERT_CAST (filter);
if (space->convert) { if (space->convert) {
videoconvert_convert_free (space->convert); videoconvert_convert_free (space->convert);
} }
/* input caps */ /* input caps */
if (!gst_video_info_from_caps (&in_info, incaps)) if (GST_VIDEO_INFO_IS_RGB (in_info)) {
goto invalid_caps;
if (in_info.finfo->flags & GST_VIDEO_FORMAT_FLAG_RGB) {
in_spec = COLOR_SPEC_RGB; in_spec = COLOR_SPEC_RGB;
} else if (in_info.finfo->flags & GST_VIDEO_FORMAT_FLAG_YUV) { } else if (GST_VIDEO_INFO_IS_YUV (in_info)) {
if (in_info.colorimetry.matrix == GST_VIDEO_COLOR_MATRIX_BT709) if (in_info->colorimetry.matrix == GST_VIDEO_COLOR_MATRIX_BT709)
in_spec = COLOR_SPEC_YUV_BT709; in_spec = COLOR_SPEC_YUV_BT709;
else else
in_spec = COLOR_SPEC_YUV_BT470_6; in_spec = COLOR_SPEC_YUV_BT470_6;
@ -257,13 +186,10 @@ gst_video_convert_set_caps (GstBaseTransform * btrans, GstCaps * incaps,
} }
/* output caps */ /* output caps */
if (!gst_video_info_from_caps (&out_info, outcaps)) if (GST_VIDEO_INFO_IS_RGB (out_info)) {
goto invalid_caps;
if (out_info.finfo->flags & GST_VIDEO_FORMAT_FLAG_RGB) {
out_spec = COLOR_SPEC_RGB; out_spec = COLOR_SPEC_RGB;
} else if (out_info.finfo->flags & GST_VIDEO_FORMAT_FLAG_YUV) { } else if (GST_VIDEO_INFO_IS_YUV (out_info)) {
if (out_info.colorimetry.matrix == GST_VIDEO_COLOR_MATRIX_BT709) if (out_info->colorimetry.matrix == GST_VIDEO_COLOR_MATRIX_BT709)
out_spec = COLOR_SPEC_YUV_BT709; out_spec = COLOR_SPEC_YUV_BT709;
else else
out_spec = COLOR_SPEC_YUV_BT470_6; out_spec = COLOR_SPEC_YUV_BT470_6;
@ -272,42 +198,39 @@ gst_video_convert_set_caps (GstBaseTransform * btrans, GstCaps * incaps,
} }
/* these must match */ /* these must match */
if (in_info.width != out_info.width || in_info.height != out_info.height || if (in_info->width != out_info->width || in_info->height != out_info->height
in_info.fps_n != out_info.fps_n || in_info.fps_d != out_info.fps_d) || in_info->fps_n != out_info->fps_n || in_info->fps_d != out_info->fps_d)
goto format_mismatch; goto format_mismatch;
/* if present, these must match too */ /* if present, these must match too */
if (in_info.par_n != out_info.par_n || in_info.par_d != out_info.par_d) if (in_info->par_n != out_info->par_n || in_info->par_d != out_info->par_d)
goto format_mismatch; goto format_mismatch;
/* if present, these must match too */ /* if present, these must match too */
if ((in_info.flags & GST_VIDEO_FLAG_INTERLACED) != if ((in_info->flags & GST_VIDEO_FLAG_INTERLACED) !=
(out_info.flags & GST_VIDEO_FLAG_INTERLACED)) (out_info->flags & GST_VIDEO_FLAG_INTERLACED))
goto format_mismatch; goto format_mismatch;
space->from_info = in_info;
space->from_spec = in_spec; space->from_spec = in_spec;
space->to_info = out_info;
space->to_spec = out_spec; space->to_spec = out_spec;
interlaced = (in_info.flags & GST_VIDEO_FLAG_INTERLACED) != 0; interlaced = (in_info->flags & GST_VIDEO_FLAG_INTERLACED) != 0;
space->convert = space->convert =
videoconvert_convert_new (GST_VIDEO_INFO_FORMAT (&out_info), out_spec, videoconvert_convert_new (GST_VIDEO_INFO_FORMAT (out_info), out_spec,
GST_VIDEO_INFO_FORMAT (&in_info), in_spec, in_info.width, in_info.height); GST_VIDEO_INFO_FORMAT (in_info), in_spec, in_info->width,
in_info->height);
if (space->convert == NULL) if (space->convert == NULL)
goto no_convert; goto no_convert;
videoconvert_convert_set_interlaced (space->convert, interlaced); videoconvert_convert_set_interlaced (space->convert, interlaced);
/* palette, only for from data */ /* palette, only for from data */
if (GST_VIDEO_INFO_FORMAT (&space->from_info) == if (GST_VIDEO_INFO_FORMAT (in_info) ==
GST_VIDEO_FORMAT_RGB8_PALETTED GST_VIDEO_FORMAT_RGB8_PALETTED
&& GST_VIDEO_INFO_FORMAT (&space->to_info) == && GST_VIDEO_INFO_FORMAT (out_info) == GST_VIDEO_FORMAT_RGB8_PALETTED) {
GST_VIDEO_FORMAT_RGB8_PALETTED) {
goto format_mismatch; goto format_mismatch;
} else if (GST_VIDEO_INFO_FORMAT (&space->from_info) == } else if (GST_VIDEO_INFO_FORMAT (in_info) == GST_VIDEO_FORMAT_RGB8_PALETTED) {
GST_VIDEO_FORMAT_RGB8_PALETTED) {
GstBuffer *palette; GstBuffer *palette;
guint32 *data; guint32 *data;
@ -324,8 +247,7 @@ gst_video_convert_set_caps (GstBaseTransform * btrans, GstCaps * incaps,
gst_buffer_unmap (palette, data, -1); gst_buffer_unmap (palette, data, -1);
gst_buffer_unref (palette); gst_buffer_unref (palette);
} else if (GST_VIDEO_INFO_FORMAT (&space->to_info) == } else if (GST_VIDEO_INFO_FORMAT (out_info) == GST_VIDEO_FORMAT_RGB8_PALETTED) {
GST_VIDEO_FORMAT_RGB8_PALETTED) {
const guint32 *palette; const guint32 *palette;
GstBuffer *p_buf; GstBuffer *p_buf;
@ -337,37 +259,25 @@ gst_video_convert_set_caps (GstBaseTransform * btrans, GstCaps * incaps,
gst_buffer_unref (p_buf); gst_buffer_unref (p_buf);
} }
GST_DEBUG ("reconfigured %d %d", GST_VIDEO_INFO_FORMAT (&space->from_info), GST_DEBUG ("reconfigured %d %d", GST_VIDEO_INFO_FORMAT (in_info),
GST_VIDEO_INFO_FORMAT (&space->to_info)); GST_VIDEO_INFO_FORMAT (out_info));
space->negotiated = TRUE;
return TRUE; return TRUE;
/* ERRORS */ /* ERRORS */
invalid_caps:
{
GST_ERROR_OBJECT (space, "invalid caps");
goto error_done;
}
format_mismatch: format_mismatch:
{ {
GST_ERROR_OBJECT (space, "input and output formats do not match"); GST_ERROR_OBJECT (space, "input and output formats do not match");
goto error_done; return FALSE;
} }
no_convert: no_convert:
{ {
GST_ERROR_OBJECT (space, "could not create converter"); GST_ERROR_OBJECT (space, "could not create converter");
goto error_done; return FALSE;
} }
invalid_palette: invalid_palette:
{ {
GST_ERROR_OBJECT (space, "invalid palette"); GST_ERROR_OBJECT (space, "invalid palette");
goto error_done;
}
error_done:
{
space->negotiated = FALSE;
return FALSE; return FALSE;
} }
} }
@ -394,6 +304,7 @@ gst_video_convert_class_init (GstVideoConvertClass * klass)
GstElementClass *gstelement_class = (GstElementClass *) klass; GstElementClass *gstelement_class = (GstElementClass *) klass;
GstBaseTransformClass *gstbasetransform_class = GstBaseTransformClass *gstbasetransform_class =
(GstBaseTransformClass *) klass; (GstBaseTransformClass *) klass;
GstVideoFilterClass *gstvideofilter_class = (GstVideoFilterClass *) klass;
gobject_class->set_property = gst_video_convert_set_property; gobject_class->set_property = gst_video_convert_set_property;
gobject_class->get_property = gst_video_convert_get_property; gobject_class->get_property = gst_video_convert_get_property;
@ -411,19 +322,14 @@ gst_video_convert_class_init (GstVideoConvertClass * klass)
gstbasetransform_class->transform_caps = gstbasetransform_class->transform_caps =
GST_DEBUG_FUNCPTR (gst_video_convert_transform_caps); GST_DEBUG_FUNCPTR (gst_video_convert_transform_caps);
gstbasetransform_class->set_caps =
GST_DEBUG_FUNCPTR (gst_video_convert_set_caps);
gstbasetransform_class->transform_size =
GST_DEBUG_FUNCPTR (gst_video_convert_transform_size);
gstbasetransform_class->propose_allocation =
GST_DEBUG_FUNCPTR (gst_video_convert_propose_allocation);
gstbasetransform_class->decide_allocation =
GST_DEBUG_FUNCPTR (gst_video_convert_decide_allocation);
gstbasetransform_class->transform =
GST_DEBUG_FUNCPTR (gst_video_convert_transform);
gstbasetransform_class->passthrough_on_same_caps = TRUE; gstbasetransform_class->passthrough_on_same_caps = TRUE;
gstvideofilter_class->set_info =
GST_DEBUG_FUNCPTR (gst_video_convert_set_info);
gstvideofilter_class->transform_frame =
GST_DEBUG_FUNCPTR (gst_video_convert_transform_frame);
g_object_class_install_property (gobject_class, PROP_DITHER, g_object_class_install_property (gobject_class, PROP_DITHER,
g_param_spec_enum ("dither", "Dither", "Apply dithering while converting", g_param_spec_enum ("dither", "Dither", "Apply dithering while converting",
dither_method_get_type (), DITHER_NONE, dither_method_get_type (), DITHER_NONE,
@ -433,7 +339,6 @@ gst_video_convert_class_init (GstVideoConvertClass * klass)
static void static void
gst_video_convert_init (GstVideoConvert * space) gst_video_convert_init (GstVideoConvert * space)
{ {
space->negotiated = FALSE;
} }
void void
@ -472,78 +377,26 @@ gst_video_convert_get_property (GObject * object, guint property_id,
} }
} }
static gboolean
gst_video_convert_transform_size (GstBaseTransform * btrans,
GstPadDirection direction, GstCaps * caps, gsize size,
GstCaps * othercaps, gsize * othersize)
{
gboolean ret = TRUE;
GstVideoInfo info;
g_assert (size);
ret = gst_video_info_from_caps (&info, othercaps);
if (ret)
*othersize = info.size;
return ret;
}
static GstFlowReturn static GstFlowReturn
gst_video_convert_transform (GstBaseTransform * btrans, GstBuffer * inbuf, gst_video_convert_transform_frame (GstVideoFilter * filter,
GstBuffer * outbuf) GstVideoFrame * in_frame, GstVideoFrame * out_frame)
{ {
GstVideoConvert *space; GstVideoConvert *space;
GstVideoFrame in_frame, out_frame;
space = GST_VIDEO_CONVERT_CAST (btrans); space = GST_VIDEO_CONVERT_CAST (filter);
GST_DEBUG ("from %s -> to %s", GST_VIDEO_INFO_NAME (&space->from_info), GST_DEBUG ("from %s -> to %s", GST_VIDEO_INFO_NAME (&filter->in_info),
GST_VIDEO_INFO_NAME (&space->to_info)); GST_VIDEO_INFO_NAME (&filter->out_info));
if (G_UNLIKELY (!space->negotiated))
goto unknown_format;
videoconvert_convert_set_dither (space->convert, space->dither); videoconvert_convert_set_dither (space->convert, space->dither);
if (!gst_video_frame_map (&in_frame, &space->from_info, inbuf, GST_MAP_READ)) videoconvert_convert_convert (space->convert, out_frame, in_frame);
goto invalid_buffer;
if (!gst_video_frame_map (&out_frame, &space->to_info, outbuf, GST_MAP_WRITE))
goto invalid_buffer;
videoconvert_convert_convert (space->convert, &out_frame, &in_frame);
gst_video_frame_unmap (&out_frame);
gst_video_frame_unmap (&in_frame);
/* baseclass copies timestamps */ /* baseclass copies timestamps */
GST_DEBUG ("from %s -> to %s done", GST_VIDEO_INFO_NAME (&space->from_info), GST_DEBUG ("from %s -> to %s done", GST_VIDEO_INFO_NAME (&filter->in_info),
GST_VIDEO_INFO_NAME (&space->to_info)); GST_VIDEO_INFO_NAME (&filter->out_info));
return GST_FLOW_OK; return GST_FLOW_OK;
/* ERRORS */
unknown_format:
{
GST_ELEMENT_ERROR (space, CORE, NOT_IMPLEMENTED, (NULL),
("attempting to convert colorspaces between unknown formats"));
return GST_FLOW_NOT_NEGOTIATED;
}
invalid_buffer:
{
GST_ELEMENT_WARNING (space, CORE, NOT_IMPLEMENTED, (NULL),
("invalid video buffer received"));
return GST_FLOW_OK;
}
#if 0
not_supported:
{
GST_ELEMENT_ERROR (space, CORE, NOT_IMPLEMENTED, (NULL),
("cannot convert between formats"));
return GST_FLOW_NOT_SUPPORTED;
}
#endif
} }
static gboolean static gboolean

View file

@ -47,10 +47,6 @@ typedef struct _GstVideoConvertClass GstVideoConvertClass;
struct _GstVideoConvert { struct _GstVideoConvert {
GstVideoFilter element; GstVideoFilter element;
GstVideoInfo from_info;
GstVideoInfo to_info;
gboolean negotiated;
ColorSpaceColorSpec from_spec; ColorSpaceColorSpec from_spec;
ColorSpaceColorSpec to_spec; ColorSpaceColorSpec to_spec;

View file

@ -186,19 +186,15 @@ static gboolean gst_video_scale_src_event (GstBaseTransform * trans,
/* base transform vmethods */ /* base transform vmethods */
static GstCaps *gst_video_scale_transform_caps (GstBaseTransform * trans, static GstCaps *gst_video_scale_transform_caps (GstBaseTransform * trans,
GstPadDirection direction, GstCaps * caps, GstCaps * filter); GstPadDirection direction, GstCaps * caps, GstCaps * filter);
static gboolean gst_video_scale_set_caps (GstBaseTransform * trans,
GstCaps * in, GstCaps * out);
static gboolean gst_video_scale_get_unit_size (GstBaseTransform * trans,
GstCaps * caps, gsize * size);
static gboolean gst_video_scale_propose_allocation (GstBaseTransform * trans,
GstQuery * query);
static gboolean gst_video_scale_decide_allocation (GstBaseTransform * trans,
GstQuery * query);
static GstFlowReturn gst_video_scale_transform (GstBaseTransform * trans,
GstBuffer * in, GstBuffer * out);
static void gst_video_scale_fixate_caps (GstBaseTransform * base, static void gst_video_scale_fixate_caps (GstBaseTransform * base,
GstPadDirection direction, GstCaps * caps, GstCaps * othercaps); GstPadDirection direction, GstCaps * caps, GstCaps * othercaps);
static gboolean gst_video_scale_set_info (GstVideoFilter * filter,
GstCaps * in, GstVideoInfo * in_info, GstCaps * out,
GstVideoInfo * out_info);
static GstFlowReturn gst_video_scale_transform_frame (GstVideoFilter * filter,
GstVideoFrame * in, GstVideoFrame * out);
static void gst_video_scale_set_property (GObject * object, guint prop_id, static void gst_video_scale_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec); const GValue * value, GParamSpec * pspec);
static void gst_video_scale_get_property (GObject * object, guint prop_id, static void gst_video_scale_get_property (GObject * object, guint prop_id,
@ -213,6 +209,7 @@ gst_video_scale_class_init (GstVideoScaleClass * klass)
GObjectClass *gobject_class = (GObjectClass *) klass; GObjectClass *gobject_class = (GObjectClass *) klass;
GstElementClass *element_class = (GstElementClass *) klass; GstElementClass *element_class = (GstElementClass *) klass;
GstBaseTransformClass *trans_class = (GstBaseTransformClass *) klass; GstBaseTransformClass *trans_class = (GstBaseTransformClass *) klass;
GstVideoFilterClass *filter_class = (GstVideoFilterClass *) klass;
gobject_class->finalize = (GObjectFinalizeFunc) gst_video_scale_finalize; gobject_class->finalize = (GObjectFinalizeFunc) gst_video_scale_finalize;
gobject_class->set_property = gst_video_scale_set_property; gobject_class->set_property = gst_video_scale_set_property;
@ -270,16 +267,12 @@ gst_video_scale_class_init (GstVideoScaleClass * klass)
trans_class->transform_caps = trans_class->transform_caps =
GST_DEBUG_FUNCPTR (gst_video_scale_transform_caps); GST_DEBUG_FUNCPTR (gst_video_scale_transform_caps);
trans_class->set_caps = GST_DEBUG_FUNCPTR (gst_video_scale_set_caps);
trans_class->get_unit_size =
GST_DEBUG_FUNCPTR (gst_video_scale_get_unit_size);
trans_class->propose_allocation =
GST_DEBUG_FUNCPTR (gst_video_scale_propose_allocation);
trans_class->decide_allocation =
GST_DEBUG_FUNCPTR (gst_video_scale_decide_allocation);
trans_class->transform = GST_DEBUG_FUNCPTR (gst_video_scale_transform);
trans_class->fixate_caps = GST_DEBUG_FUNCPTR (gst_video_scale_fixate_caps); trans_class->fixate_caps = GST_DEBUG_FUNCPTR (gst_video_scale_fixate_caps);
trans_class->src_event = GST_DEBUG_FUNCPTR (gst_video_scale_src_event); trans_class->src_event = GST_DEBUG_FUNCPTR (gst_video_scale_src_event);
filter_class->set_info = GST_DEBUG_FUNCPTR (gst_video_scale_set_info);
filter_class->transform_frame =
GST_DEBUG_FUNCPTR (gst_video_scale_transform_frame);
} }
static void static void
@ -450,92 +443,21 @@ gst_video_scale_transform_caps (GstBaseTransform * trans,
return ret; return ret;
} }
/* Answer the allocation query downstream. This is only called for
* non-passthrough cases */
static gboolean static gboolean
gst_video_scale_propose_allocation (GstBaseTransform * trans, GstQuery * query) gst_video_scale_set_info (GstVideoFilter * filter, GstCaps * in,
GstVideoInfo * in_info, GstCaps * out, GstVideoInfo * out_info)
{ {
GstVideoScale *scale = GST_VIDEO_SCALE_CAST (trans); GstVideoScale *videoscale = GST_VIDEO_SCALE (filter);
GstBufferPool *pool;
GstCaps *caps;
gboolean need_pool;
guint size;
gst_query_parse_allocation (query, &caps, &need_pool);
size = GST_VIDEO_INFO_SIZE (&scale->from_info);
if (need_pool) {
GstStructure *structure;
pool = gst_video_buffer_pool_new ();
structure = gst_buffer_pool_get_config (pool);
gst_buffer_pool_config_set (structure, caps, size, 0, 0, 0, 15);
if (!gst_buffer_pool_set_config (pool, structure))
goto config_failed;
} else
pool = NULL;
gst_query_set_allocation_params (query, size, 0, 0, 0, 15, pool);
gst_object_unref (pool);
gst_query_add_allocation_meta (query, GST_VIDEO_META_API);
return TRUE;
/* ERRORS */
config_failed:
{
GST_ERROR_OBJECT (scale, "failed to set config.");
gst_object_unref (pool);
return FALSE;
}
}
/* configure the allocation query that was answered downstream, we can configure
* some properties on it. Only called in passthrough mode. */
static gboolean
gst_video_scale_decide_allocation (GstBaseTransform * trans, GstQuery * query)
{
GstBufferPool *pool = NULL;
guint size, min, max, prefix, alignment;
gst_query_parse_allocation_params (query, &size, &min, &max, &prefix,
&alignment, &pool);
if (pool) {
GstStructure *config;
config = gst_buffer_pool_get_config (pool);
gst_buffer_pool_config_add_option (config,
GST_BUFFER_POOL_OPTION_VIDEO_META);
gst_buffer_pool_set_config (pool, config);
}
return TRUE;
}
static gboolean
gst_video_scale_set_caps (GstBaseTransform * trans, GstCaps * in, GstCaps * out)
{
GstVideoScale *videoscale = GST_VIDEO_SCALE (trans);
gboolean ret;
GstVideoInfo in_info, out_info;
gint from_dar_n, from_dar_d, to_dar_n, to_dar_d; gint from_dar_n, from_dar_d, to_dar_n, to_dar_d;
ret = gst_video_info_from_caps (&in_info, in); if (!gst_util_fraction_multiply (in_info->width,
ret &= gst_video_info_from_caps (&out_info, out); in_info->height, out_info->par_n, out_info->par_d, &from_dar_n,
if (!ret)
goto invalid_formats;
if (!gst_util_fraction_multiply (in_info.width,
in_info.height, out_info.par_n, out_info.par_d, &from_dar_n,
&from_dar_d)) { &from_dar_d)) {
from_dar_n = from_dar_d = -1; from_dar_n = from_dar_d = -1;
} }
if (!gst_util_fraction_multiply (out_info.width, if (!gst_util_fraction_multiply (out_info->width,
out_info.height, out_info.par_n, out_info.par_d, &to_dar_n, out_info->height, out_info->par_n, out_info->par_d, &to_dar_n,
&to_dar_d)) { &to_dar_d)) {
to_dar_n = to_dar_d = -1; to_dar_n = to_dar_d = -1;
} }
@ -546,17 +468,17 @@ gst_video_scale_set_caps (GstBaseTransform * trans, GstCaps * in, GstCaps * out)
gint n, d, to_h, to_w; gint n, d, to_h, to_w;
if (from_dar_n != -1 && from_dar_d != -1 if (from_dar_n != -1 && from_dar_d != -1
&& gst_util_fraction_multiply (from_dar_n, from_dar_d, out_info.par_n, && gst_util_fraction_multiply (from_dar_n, from_dar_d,
out_info.par_d, &n, &d)) { out_info->par_n, out_info->par_d, &n, &d)) {
to_h = gst_util_uint64_scale_int (out_info.width, d, n); to_h = gst_util_uint64_scale_int (out_info->width, d, n);
if (to_h <= out_info.height) { if (to_h <= out_info->height) {
videoscale->borders_h = out_info.height - to_h; videoscale->borders_h = out_info->height - to_h;
videoscale->borders_w = 0; videoscale->borders_w = 0;
} else { } else {
to_w = gst_util_uint64_scale_int (out_info.height, n, d); to_w = gst_util_uint64_scale_int (out_info->height, n, d);
g_assert (to_w <= out_info.width); g_assert (to_w <= out_info->width);
videoscale->borders_h = 0; videoscale->borders_h = 0;
videoscale->borders_w = out_info.width - to_w; videoscale->borders_w = out_info->width - to_w;
} }
} else { } else {
GST_WARNING_OBJECT (videoscale, "Can't calculate borders"); GST_WARNING_OBJECT (videoscale, "Can't calculate borders");
@ -568,42 +490,19 @@ gst_video_scale_set_caps (GstBaseTransform * trans, GstCaps * in, GstCaps * out)
if (videoscale->tmp_buf) if (videoscale->tmp_buf)
g_free (videoscale->tmp_buf); g_free (videoscale->tmp_buf);
videoscale->tmp_buf = g_malloc (out_info.width * 8 * 4); videoscale->tmp_buf = g_malloc (out_info->width * 8 * 4);
gst_base_transform_set_passthrough (trans, gst_base_transform_set_passthrough (GST_BASE_TRANSFORM (filter),
(in_info.width == out_info.width && in_info.height == out_info.height)); (in_info->width == out_info->width
&& in_info->height == out_info->height));
GST_DEBUG_OBJECT (videoscale, "from=%dx%d (par=%d/%d dar=%d/%d), size %" GST_DEBUG_OBJECT (videoscale, "from=%dx%d (par=%d/%d dar=%d/%d), size %"
G_GSIZE_FORMAT " -> to=%dx%d (par=%d/%d dar=%d/%d borders=%d:%d), " G_GSIZE_FORMAT " -> to=%dx%d (par=%d/%d dar=%d/%d borders=%d:%d), "
"size %" G_GSIZE_FORMAT, "size %" G_GSIZE_FORMAT,
in_info.width, in_info.height, out_info.par_n, out_info.par_d, in_info->width, in_info->height, out_info->par_n, out_info->par_d,
from_dar_n, from_dar_d, in_info.size, out_info.width, from_dar_n, from_dar_d, in_info->size, out_info->width,
out_info.height, out_info.par_n, out_info.par_d, to_dar_n, to_dar_d, out_info->height, out_info->par_n, out_info->par_d, to_dar_n, to_dar_d,
videoscale->borders_w, videoscale->borders_h, out_info.size); videoscale->borders_w, videoscale->borders_h, out_info->size);
videoscale->from_info = in_info;
videoscale->to_info = out_info;
return TRUE;
/* ERRORS */
invalid_formats:
{
GST_DEBUG_OBJECT (videoscale, "could not parse formats");
return FALSE;
}
}
static gboolean
gst_video_scale_get_unit_size (GstBaseTransform * trans, GstCaps * caps,
gsize * size)
{
GstVideoInfo info;
if (!gst_video_info_from_caps (&info, caps))
return FALSE;
*size = info.size;
return TRUE; return TRUE;
} }
@ -1149,12 +1048,11 @@ _get_black_for_format (GstVideoFormat format)
} }
static GstFlowReturn static GstFlowReturn
gst_video_scale_transform (GstBaseTransform * trans, GstBuffer * in, gst_video_scale_transform_frame (GstVideoFilter * filter,
GstBuffer * out) GstVideoFrame * in_frame, GstVideoFrame * out_frame)
{ {
GstVideoScale *videoscale = GST_VIDEO_SCALE (trans); GstVideoScale *videoscale = GST_VIDEO_SCALE (filter);
GstFlowReturn ret = GST_FLOW_OK; GstFlowReturn ret = GST_FLOW_OK;
GstVideoFrame in_frame, out_frame;
VSImage dest[4] = { {NULL,}, }; VSImage dest[4] = { {NULL,}, };
VSImage src[4] = { {NULL,}, }; VSImage src[4] = { {NULL,}, };
gint method; gint method;
@ -1168,23 +1066,20 @@ gst_video_scale_transform (GstBaseTransform * trans, GstBuffer * in,
add_borders = videoscale->add_borders; add_borders = videoscale->add_borders;
GST_OBJECT_UNLOCK (videoscale); GST_OBJECT_UNLOCK (videoscale);
format = GST_VIDEO_INFO_FORMAT (&videoscale->from_info); format = GST_VIDEO_INFO_FORMAT (&filter->in_info);
black = _get_black_for_format (format); black = _get_black_for_format (format);
if (videoscale->from_info.width == 1) { if (filter->in_info.width == 1) {
method = GST_VIDEO_SCALE_NEAREST; method = GST_VIDEO_SCALE_NEAREST;
} }
if (method == GST_VIDEO_SCALE_4TAP && if (method == GST_VIDEO_SCALE_4TAP &&
(videoscale->from_info.width < 4 || videoscale->from_info.height < 4)) { (filter->in_info.width < 4 || filter->in_info.height < 4)) {
method = GST_VIDEO_SCALE_BILINEAR; method = GST_VIDEO_SCALE_BILINEAR;
} }
gst_video_frame_map (&in_frame, &videoscale->from_info, in, GST_MAP_READ); for (i = 0; i < GST_VIDEO_FRAME_N_PLANES (in_frame); i++) {
gst_video_frame_map (&out_frame, &videoscale->to_info, out, GST_MAP_WRITE); gst_video_scale_setup_vs_image (&src[i], in_frame, i, 0, 0);
gst_video_scale_setup_vs_image (&dest[i], out_frame, i,
for (i = 0; i < GST_VIDEO_FRAME_N_PLANES (&in_frame); i++) {
gst_video_scale_setup_vs_image (&src[i], &in_frame, i, 0, 0);
gst_video_scale_setup_vs_image (&dest[i], &out_frame, i,
videoscale->borders_w, videoscale->borders_h); videoscale->borders_w, videoscale->borders_h);
} }
@ -1410,13 +1305,6 @@ gst_video_scale_transform (GstBaseTransform * trans, GstBuffer * in,
goto unsupported; goto unsupported;
} }
GST_LOG_OBJECT (videoscale, "pushing buffer of %" G_GSIZE_FORMAT " bytes",
gst_buffer_get_size (out));
done:
gst_video_frame_unmap (&out_frame);
gst_video_frame_unmap (&in_frame);
return ret; return ret;
/* ERRORS */ /* ERRORS */
@ -1424,22 +1312,21 @@ unsupported:
{ {
GST_ELEMENT_ERROR (videoscale, STREAM, NOT_IMPLEMENTED, (NULL), GST_ELEMENT_ERROR (videoscale, STREAM, NOT_IMPLEMENTED, (NULL),
("Unsupported format %d for scaling method %d", format, method)); ("Unsupported format %d for scaling method %d", format, method));
ret = GST_FLOW_ERROR; return GST_FLOW_ERROR;
goto done;
} }
unknown_mode: unknown_mode:
{ {
GST_ELEMENT_ERROR (videoscale, STREAM, NOT_IMPLEMENTED, (NULL), GST_ELEMENT_ERROR (videoscale, STREAM, NOT_IMPLEMENTED, (NULL),
("Unknown scaling method %d", videoscale->method)); ("Unknown scaling method %d", videoscale->method));
ret = GST_FLOW_ERROR; return GST_FLOW_ERROR;
goto done;
} }
} }
static gboolean static gboolean
gst_video_scale_src_event (GstBaseTransform * trans, GstEvent * event) gst_video_scale_src_event (GstBaseTransform * trans, GstEvent * event)
{ {
GstVideoScale *videoscale = GST_VIDEO_SCALE (trans); GstVideoScale *videoscale = GST_VIDEO_SCALE_CAST (trans);
GstVideoFilter *filter = GST_VIDEO_FILTER_CAST (trans);
gboolean ret; gboolean ret;
gdouble a; gdouble a;
GstStructure *structure; GstStructure *structure;
@ -1455,12 +1342,11 @@ gst_video_scale_src_event (GstBaseTransform * trans, GstEvent * event)
structure = (GstStructure *) gst_event_get_structure (event); structure = (GstStructure *) gst_event_get_structure (event);
if (gst_structure_get_double (structure, "pointer_x", &a)) { if (gst_structure_get_double (structure, "pointer_x", &a)) {
gst_structure_set (structure, "pointer_x", G_TYPE_DOUBLE, gst_structure_set (structure, "pointer_x", G_TYPE_DOUBLE,
a * videoscale->from_info.width / videoscale->to_info.width, NULL); a * filter->in_info.width / filter->out_info.width, NULL);
} }
if (gst_structure_get_double (structure, "pointer_y", &a)) { if (gst_structure_get_double (structure, "pointer_y", &a)) {
gst_structure_set (structure, "pointer_y", G_TYPE_DOUBLE, gst_structure_set (structure, "pointer_y", G_TYPE_DOUBLE,
a * videoscale->from_info.height / videoscale->to_info.height, a * filter->in_info.height / filter->out_info.height, NULL);
NULL);
} }
break; break;
default: default:

View file

@ -80,10 +80,6 @@ struct _GstVideoScale {
int submethod; int submethod;
double envelope; double envelope;
/* negotiated stuff */
GstVideoInfo from_info;
GstVideoInfo to_info;
gint borders_h; gint borders_h;
gint borders_w; gint borders_w;