gstreamer/tests/check/elements/rawvideoparse.c
Carlos Rafael Giani 232902369c rawparse: Fix and extend unit tests
* Add caps change test to unit tests
* Cleanup leftover buffers after each unit test
* Add missing rawvideoparse entry in .gitignore

https://bugzilla.gnome.org/show_bug.cgi?id=769637
2016-08-18 10:53:58 +03:00

628 lines
21 KiB
C

/* GStreamer
*
* unit test for rawvideoparse
*
* Copyright (C) <2016> Carlos Rafael Giani <dv at pseudoterminal dot org>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
* Boston, MA 02110-1301, USA.
*/
/* FIXME: GValueArray is deprecated, but there is currently no viabla alternative
* See https://bugzilla.gnome.org/show_bug.cgi?id=667228 */
#define GLIB_DISABLE_DEPRECATION_WARNINGS
#include <gst/check/gstcheck.h>
#include <gst/video/video.h>
/* The checks use as test data an 8x8 Y444 image, with 25 Hz framerate. In the
* sink caps configuration, the stride is 8 bytes, and the frames are tightly
* packed together. In the properties configuration, the stride is 10 bytes, the
* planes aren't tightly packed (there are 20 bytes between the planes), and the
* frames overall have padding between them (the overall frame size is
* stride (10) * height (8) * num-planes (3) + bytes-between-planes (20) * 2
* = 280 bytes, and the frame stride is 500 bytes, so there are 220 bytes of
* extra padding between frames).
*
* In the test 8x8 frame, the pixels are all set to #000000, except for two
* pixels: (xofs+1 yofs+0) is set to #8899AA, (xofs+0 yofs+1) is set to #112233.
* The first frame uses the offsets xofs=0 yofs=0. The second frame uses
* xofs=1 yofs=0 etc. For each configuration, there is a separate set of frames,
* each stored in the GstAdapter in the Context struct.
*
* During the tests, as part of the checks, the pixels are verified to have the
* right values. The pattern of the pixels was chosen to easily detect stride
* errors, incorrect plane offsets etc.
*/
#define TEST_WIDTH 8
#define TEST_HEIGHT 8
#define TEST_FRAMERATE_N 25
#define TEST_FRAMERATE_D 1
#define TEST_FRAME_FORMAT GST_VIDEO_FORMAT_Y444
#define NUM_TEST_PLANES 3
#define PROP_CTX_PLANE_STRIDE 10
#define PROP_CTX_FRAME_STRIDE 500
#define PROP_CTX_PLANE_PADDING 20
#define PROP_CTX_PLANE_SIZE (PROP_CTX_PLANE_STRIDE * TEST_HEIGHT + PROP_CTX_PLANE_PADDING)
GstElement *rawvideoparse;
/* For ease of programming we use globals to keep refs for our floating
* src and sink pads we create; otherwise we always have to do get_pad,
* get_peer, and then remove references in every test function */
static GstPad *mysrcpad, *mysinkpad;
typedef struct
{
GstAdapter *data;
guint plane_stride;
guint plane_size;
}
Context;
static Context properties_ctx, sinkcaps_ctx;
static void
set_pixel (Context const *ctx, guint8 * pixels, guint x, guint y, guint32 color)
{
guint i;
guint ofs = y * ctx->plane_stride + x;
for (i = 0; i < NUM_TEST_PLANES; ++i)
pixels[ctx->plane_size * i + ofs] =
(color >> ((NUM_TEST_PLANES - 1 - i) * 8)) & 0xFF;
}
static guint32
get_pixel (Context const *ctx, const guint8 * pixels, guint x, guint y)
{
guint i;
guint ofs = y * ctx->plane_stride + x;
guint32 color = 0;
for (i = 0; i < NUM_TEST_PLANES; ++i)
color |=
((guint32) (pixels[ctx->plane_size * i + ofs])) << ((NUM_TEST_PLANES -
1 - i) * 8);
return color;
}
static void
fill_test_pattern (Context const *ctx, GstBuffer * buffer, guint xofs,
guint yofs)
{
guint8 *pixels;
GstMapInfo map_info;
gst_buffer_map (buffer, &map_info, GST_MAP_WRITE);
pixels = map_info.data;
memset (pixels, 0, ctx->plane_size * NUM_TEST_PLANES);
set_pixel (ctx, pixels, 1 + xofs, 0 + yofs, 0x8899AA);
set_pixel (ctx, pixels, 0 + xofs, 1 + yofs, 0x112233);
gst_buffer_unmap (buffer, &map_info);
}
static void
check_test_pattern (Context const *ctx, GstBuffer * buffer, guint xofs,
guint yofs)
{
guint x, y;
guint8 *pixels;
GstMapInfo map_info;
gst_buffer_map (buffer, &map_info, GST_MAP_READ);
pixels = map_info.data;
fail_unless_equals_uint64_hex (get_pixel (ctx, pixels, 1 + xofs, 0 + yofs),
0x8899AA);
fail_unless_equals_uint64_hex (get_pixel (ctx, pixels, 0 + xofs, 1 + yofs),
0x112233);
for (y = 0; y < TEST_HEIGHT; ++y) {
for (x = 0; x < TEST_WIDTH; ++x) {
if ((x == (1 + xofs) && y == (0 + yofs)) || (x == (0 + xofs)
&& y == (1 + yofs)))
continue;
fail_unless_equals_uint64_hex (get_pixel (ctx, pixels, x, y), 0x000000);
}
}
gst_buffer_unmap (buffer, &map_info);
}
static void
setup_rawvideoparse (gboolean use_sink_caps,
gboolean set_properties, GstCaps * incaps, GstFormat format)
{
guint i;
/* Setup the rawvideoparse element and the pads */
static GstStaticPadTemplate sinktemplate = GST_STATIC_PAD_TEMPLATE ("sink",
GST_PAD_SINK,
GST_PAD_ALWAYS,
GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE (GST_VIDEO_FORMATS_ALL))
);
static GstStaticPadTemplate srctemplate = GST_STATIC_PAD_TEMPLATE ("src",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS_ANY);
rawvideoparse = gst_check_setup_element ("rawvideoparse");
properties_ctx.plane_stride = PROP_CTX_PLANE_STRIDE;
properties_ctx.plane_size = PROP_CTX_PLANE_SIZE;
properties_ctx.data = gst_adapter_new ();
sinkcaps_ctx.plane_stride = TEST_WIDTH;
sinkcaps_ctx.plane_size = TEST_WIDTH * TEST_HEIGHT;
sinkcaps_ctx.data = gst_adapter_new ();
g_object_set (G_OBJECT (rawvideoparse), "use-sink-caps", use_sink_caps, NULL);
if (set_properties) {
GValueArray *plane_offsets, *plane_strides;
GValue val = G_VALUE_INIT;
g_value_init (&val, G_TYPE_UINT);
plane_offsets = g_value_array_new (NUM_TEST_PLANES);
for (i = 0; i < NUM_TEST_PLANES; ++i) {
g_value_set_uint (&val, properties_ctx.plane_size * i);
g_value_array_insert (plane_offsets, i, &val);
}
plane_strides = g_value_array_new (NUM_TEST_PLANES);
for (i = 0; i < NUM_TEST_PLANES; ++i) {
g_value_set_uint (&val, properties_ctx.plane_stride);
g_value_array_insert (plane_strides, i, &val);
}
g_value_unset (&val);
g_object_set (G_OBJECT (rawvideoparse), "width", TEST_WIDTH, "height",
TEST_HEIGHT, "frame-stride", PROP_CTX_FRAME_STRIDE, "framerate",
TEST_FRAMERATE_N, TEST_FRAMERATE_D, "plane-offsets", plane_offsets,
"plane-strides", plane_strides, "format", TEST_FRAME_FORMAT, NULL);
g_value_array_free (plane_offsets);
g_value_array_free (plane_strides);
}
/* Check that the plane stride/offset values are correct */
{
GValueArray *plane_offsets_array;
GValueArray *plane_strides_array;
/* By default, 320x240 i420 is used as format */
guint plane_offsets[3] = { 0, 76800, 96000 };
guint plane_strides[3] = { 320, 160, 160 };
if (set_properties) {
/* When properties are explicitely set, we use Y444 as video format,
* so in that case, plane stride values are all the same */
plane_offsets[0] = properties_ctx.plane_size * 0;
plane_offsets[1] = properties_ctx.plane_size * 1;
plane_offsets[2] = properties_ctx.plane_size * 2;
plane_strides[0] = plane_strides[1] = plane_strides[2] =
properties_ctx.plane_stride;
}
g_object_get (G_OBJECT (rawvideoparse), "plane-offsets",
&plane_offsets_array, "plane-strides", &plane_strides_array, NULL);
fail_unless (plane_offsets_array != NULL);
fail_unless (plane_strides_array != NULL);
fail_unless (plane_offsets_array->n_values ==
plane_strides_array->n_values);
for (i = 0; i < plane_offsets_array->n_values; ++i) {
GValue *gvalue;
gvalue = g_value_array_get_nth (plane_offsets_array, i);
fail_unless (gvalue != NULL);
fail_unless_equals_uint64 (plane_offsets[i], g_value_get_uint (gvalue));
gvalue = g_value_array_get_nth (plane_strides_array, i);
fail_unless (gvalue != NULL);
fail_unless_equals_uint64 (plane_strides[i], g_value_get_uint (gvalue));
}
g_value_array_free (plane_offsets_array);
g_value_array_free (plane_strides_array);
}
fail_unless (gst_element_set_state (rawvideoparse,
GST_STATE_PAUSED) == GST_STATE_CHANGE_SUCCESS,
"could not set to paused");
mysrcpad = gst_check_setup_src_pad (rawvideoparse, &srctemplate);
mysinkpad = gst_check_setup_sink_pad (rawvideoparse, &sinktemplate);
gst_pad_set_active (mysrcpad, TRUE);
gst_pad_set_active (mysinkpad, TRUE);
gst_check_setup_events (mysrcpad, rawvideoparse, incaps, format);
if (incaps)
gst_caps_unref (incaps);
/* Fill the adapters with test frames */
for (i = 0; i < 10; ++i) {
GstBuffer *buffer =
gst_buffer_new_allocate (NULL, PROP_CTX_FRAME_STRIDE, NULL);
gst_buffer_memset (buffer, 0, 0xCC, gst_buffer_get_size (buffer));
fill_test_pattern (&properties_ctx, buffer, i, 0);
gst_adapter_push (properties_ctx.data, buffer);
}
for (i = 0; i < 10; ++i) {
GstBuffer *buffer =
gst_buffer_new_allocate (NULL, sinkcaps_ctx.plane_size * 3, NULL);
gst_buffer_memset (buffer, 0, 0xCC, gst_buffer_get_size (buffer));
fill_test_pattern (&sinkcaps_ctx, buffer, i, 0);
gst_adapter_push (sinkcaps_ctx.data, buffer);
}
}
static void
cleanup_rawvideoparse (void)
{
int num_buffers, i;
gst_pad_set_active (mysrcpad, FALSE);
gst_pad_set_active (mysinkpad, FALSE);
gst_check_teardown_src_pad (rawvideoparse);
gst_check_teardown_sink_pad (rawvideoparse);
gst_check_teardown_element (rawvideoparse);
g_object_unref (G_OBJECT (properties_ctx.data));
g_object_unref (G_OBJECT (sinkcaps_ctx.data));
if (buffers != NULL) {
num_buffers = g_list_length (buffers);
for (i = 0; i < num_buffers; ++i) {
GstBuffer *buf = GST_BUFFER (buffers->data);
buffers = g_list_remove (buffers, buf);
gst_buffer_unref (buf);
}
g_list_free (buffers);
buffers = NULL;
}
}
static void
push_data_and_check_output (Context * ctx, gsize num_in_bytes,
gsize expected_num_out_bytes, gint64 expected_pts, gint64 expected_dur,
guint expected_num_buffers_in_list, guint buf_idx, guint xofs, guint yofs)
{
GstBuffer *inbuf, *outbuf;
guint num_buffers;
/* Simulate upstream input by taking num_in_bytes bytes from the adapter */
inbuf = gst_adapter_take_buffer (ctx->data, num_in_bytes);
fail_unless (inbuf != NULL);
/* Push the input data and check that the output buffers list grew as
* expected */
fail_unless (gst_pad_push (mysrcpad, inbuf) == GST_FLOW_OK);
num_buffers = g_list_length (buffers);
fail_unless_equals_int (num_buffers, expected_num_buffers_in_list);
/* Take the output buffer */
outbuf = g_list_nth_data (buffers, buf_idx);
fail_unless (outbuf != NULL);
/* Verify size, PTS, duration of the output buffer */
fail_unless_equals_uint64 (expected_num_out_bytes,
gst_buffer_get_size (outbuf));
fail_unless_equals_uint64 (expected_pts, GST_BUFFER_PTS (outbuf));
fail_unless_equals_uint64 (expected_dur, GST_BUFFER_DURATION (outbuf));
/* Check that the pixels have the correct values */
check_test_pattern (ctx, outbuf, xofs, yofs);
}
GST_START_TEST (test_push_unaligned_data_properties_config)
{
setup_rawvideoparse (FALSE, TRUE, NULL, GST_FORMAT_BYTES);
/* Send in data buffers that are not aligned to multiples of the
* frame size (= sample size * num_channels). This tests if rawvideoparse
* aligns output data properly.
*
* The second line sends a buffer with multiple frames inside.
* rawvideoparse will then parse this buffer repeatedly (and prepend
* leftover data from the earlier parse iteration), explaining why
* all of a sudden there are 4 output buffers, compared to just one
* earlier. The output data is expected to be 280 bytes large, since this
* is the size of the actual frame, without extra padding at the end.
*/
push_data_and_check_output (&properties_ctx, 511, 280, GST_MSECOND * 0,
GST_MSECOND * 40, 1, 0, 0, 0);
push_data_and_check_output (&properties_ctx, 1940, 280, GST_MSECOND * 40,
GST_MSECOND * 40, 4, 1, 1, 0);
push_data_and_check_output (&properties_ctx, 10, 280, GST_MSECOND * 80,
GST_MSECOND * 40, 4, 2, 2, 0);
cleanup_rawvideoparse ();
}
GST_END_TEST;
GST_START_TEST (test_push_unaligned_data_sink_caps_config)
{
GstVideoInfo vinfo;
GstCaps *caps;
/* This test is essentially the same as test_push_unaligned_data_properties_config,
* except that rawvideoparse uses the sink caps config instead of the property config.
* Also, the input sizes are different, since the sink caps config does not use extra
* padding between planes and does use a stride that directly corresponds to the width,
* resulting in smaller frame size (192 bytes vs 280 bytes). */
gst_video_info_set_format (&vinfo, TEST_FRAME_FORMAT, TEST_WIDTH,
TEST_HEIGHT);
GST_VIDEO_INFO_FPS_N (&vinfo) = 25;
GST_VIDEO_INFO_FPS_D (&vinfo) = 1;
caps = gst_video_info_to_caps (&vinfo);
setup_rawvideoparse (TRUE, FALSE, caps, GST_FORMAT_BYTES);
push_data_and_check_output (&sinkcaps_ctx, 250, 192, GST_MSECOND * 0,
GST_MSECOND * 40, 1, 0, 0, 0);
push_data_and_check_output (&sinkcaps_ctx, 811, 192, GST_MSECOND * 40,
GST_MSECOND * 40, 5, 1, 1, 0);
push_data_and_check_output (&sinkcaps_ctx, 10, 192, GST_MSECOND * 80,
GST_MSECOND * 40, 5, 2, 2, 0);
cleanup_rawvideoparse ();
}
GST_END_TEST;
GST_START_TEST (test_config_switch)
{
GstVideoInfo vinfo;
GstCaps *caps;
/* Start processing with the properties config active, then mid-stream switch to
* the sink caps config. Since the sink caps config does not use padding, its
* frame size is smaller. The buffer duration stays the same (since it only depends
* on the framerate), but the expected output buffer size is different). */
gst_video_info_set_format (&vinfo, TEST_FRAME_FORMAT, TEST_WIDTH,
TEST_HEIGHT);
GST_VIDEO_INFO_FPS_N (&vinfo) = 25;
GST_VIDEO_INFO_FPS_D (&vinfo) = 1;
caps = gst_video_info_to_caps (&vinfo);
setup_rawvideoparse (FALSE, TRUE, caps, GST_FORMAT_BYTES);
/* Push in data with properties config active */
push_data_and_check_output (&properties_ctx, 500, 280, GST_MSECOND * 0,
GST_MSECOND * 40, 1, 0, 0, 0);
push_data_and_check_output (&properties_ctx, 500, 280, GST_MSECOND * 40,
GST_MSECOND * 40, 2, 1, 1, 0);
/* Perform the switch */
g_object_set (G_OBJECT (rawvideoparse), "use-sink-caps", TRUE, NULL);
/* Push in data with sink caps config active, expecting a different frame size */
push_data_and_check_output (&sinkcaps_ctx, 192, 192, GST_MSECOND * 80,
GST_MSECOND * 40, 3, 2, 0, 0);
cleanup_rawvideoparse ();
}
GST_END_TEST;
GST_START_TEST (test_push_with_no_framerate)
{
/* Test the special case when no framerate is set. The parser is expected to
* still work then, but without setting duration or PTS/DTS (it cannot do that,
* because these require a nonzero framerate). The output buffers have PTS 0,
* all subsequent ones have no set PTS. */
setup_rawvideoparse (FALSE, TRUE, NULL, GST_FORMAT_BYTES);
g_object_set (G_OBJECT (rawvideoparse), "framerate", 0, 1, NULL);
push_data_and_check_output (&properties_ctx, 500, 280, 0, GST_CLOCK_TIME_NONE,
1, 0, 0, 0);
push_data_and_check_output (&properties_ctx, 500, 280, GST_CLOCK_TIME_NONE,
GST_CLOCK_TIME_NONE, 2, 1, 1, 0);
cleanup_rawvideoparse ();
}
GST_END_TEST;
GST_START_TEST (test_computed_plane_strides)
{
/* Test how plane strides & offsets are (re)computed if custom offsets/strides
* are disabled, and how they are preserved if they are enabled. */
GValueArray *plane_offsets_array;
GValueArray *plane_strides_array;
guint i;
guint const expected_comp_psize = TEST_WIDTH * TEST_HEIGHT;
setup_rawvideoparse (FALSE, TRUE, NULL, GST_FORMAT_BYTES);
/* The setup set a custom set of plane offsets and strides together with
* width=TEST_WIDTH and height=TEST_HEIGHT. Check that the offsets & strides
* are preserved even after setting new, different width & height values. */
g_object_set (G_OBJECT (rawvideoparse), "width", TEST_WIDTH * 2,
"height", TEST_HEIGHT * 2, NULL);
g_object_get (G_OBJECT (rawvideoparse), "plane-offsets",
&plane_offsets_array, "plane-strides", &plane_strides_array, NULL);
for (i = 0; i < plane_offsets_array->n_values; ++i) {
GValue *gvalue;
/* See setup_rawvideoparse() for how the offsets & strides are defined
* there. Offsets are set to plane_size*plane_index, and strides are
* set to the properties_ctx.plane_stride value. */
gvalue = g_value_array_get_nth (plane_offsets_array, i);
fail_unless (gvalue != NULL);
fail_unless_equals_uint64 (properties_ctx.plane_size * i,
g_value_get_uint (gvalue));
gvalue = g_value_array_get_nth (plane_strides_array, i);
fail_unless (gvalue != NULL);
fail_unless_equals_uint64 (properties_ctx.plane_stride,
g_value_get_uint (gvalue));
}
g_value_array_free (plane_offsets_array);
g_value_array_free (plane_strides_array);
/* Discard the custom planes&offsets, re-enabling computed values. */
g_object_set (G_OBJECT (rawvideoparse), "plane-offsets", (GValueArray *) NULL,
"plane-strides", (GValueArray *) NULL, NULL);
/* The strides & offsets should have been recomputed by now. Since the Y444
* format is used, all strides are the same, and should equal the frame width
* (which was set to TEST_WIDTH*2 earlier). Plane offsets should be
* plane_size*plane_index, with plane_size set to (TEST_WIDTH*2 * TEST_HEIGHT*2),
* or TEST_WIDTH*TEST_HEIGHT*4 (-> expected_comp_psize*4). */
g_object_get (G_OBJECT (rawvideoparse), "plane-offsets",
&plane_offsets_array, "plane-strides", &plane_strides_array, NULL);
for (i = 0; i < plane_offsets_array->n_values; ++i) {
GValue *gvalue;
gvalue = g_value_array_get_nth (plane_offsets_array, i);
fail_unless (gvalue != NULL);
fail_unless_equals_uint64 (expected_comp_psize * 4 * i,
g_value_get_uint (gvalue));
gvalue = g_value_array_get_nth (plane_strides_array, i);
fail_unless (gvalue != NULL);
fail_unless_equals_uint64 (TEST_WIDTH * 2, g_value_get_uint (gvalue));
}
g_value_array_free (plane_offsets_array);
g_value_array_free (plane_strides_array);
/* Again change the width & height values. width=TEST_WIDTH, height=TEST_HEIGHT.
* However, this time, offsets&strides are computed; the current values should
* not be preserved. Expected plane stride and offset values are similar to
* above, expect that no multiplications by 2 are present (since the TEST_WIDTH
* and TEST_HEIGHT values were passed without multiplying them). */
g_object_set (G_OBJECT (rawvideoparse), "width", TEST_WIDTH,
"height", TEST_HEIGHT, NULL);
g_object_get (G_OBJECT (rawvideoparse), "plane-offsets",
&plane_offsets_array, "plane-strides", &plane_strides_array, NULL);
for (i = 0; i < plane_offsets_array->n_values; ++i) {
GValue *gvalue;
gvalue = g_value_array_get_nth (plane_offsets_array, i);
fail_unless (gvalue != NULL);
fail_unless_equals_uint64 (expected_comp_psize * i,
g_value_get_uint (gvalue));
gvalue = g_value_array_get_nth (plane_strides_array, i);
fail_unless (gvalue != NULL);
fail_unless_equals_uint64 (TEST_WIDTH, g_value_get_uint (gvalue));
}
g_value_array_free (plane_offsets_array);
g_value_array_free (plane_strides_array);
cleanup_rawvideoparse ();
}
GST_END_TEST;
GST_START_TEST (test_change_caps)
{
GstVideoInfo vinfo;
GstCaps *caps;
/* Start processing with the sink caps config active, using the
* default width/height/format and 25 Hz frame rate for the caps.
* Push some data, then change caps (25 Hz -> 50 Hz).
* Check that the changed caps are handled properly. */
gst_video_info_set_format (&vinfo, TEST_FRAME_FORMAT, TEST_WIDTH,
TEST_HEIGHT);
GST_VIDEO_INFO_FPS_N (&vinfo) = 25;
GST_VIDEO_INFO_FPS_D (&vinfo) = 1;
caps = gst_video_info_to_caps (&vinfo);
setup_rawvideoparse (TRUE, FALSE, caps, GST_FORMAT_BYTES);
/* Push in data with sink config active, expecting duration calculations
* to be based on the 25 Hz frame rate */
push_data_and_check_output (&sinkcaps_ctx, 192, 192, GST_MSECOND * 0,
GST_MSECOND * 40, 1, 0, 0, 0);
push_data_and_check_output (&sinkcaps_ctx, 192, 192, GST_MSECOND * 40,
GST_MSECOND * 40, 2, 1, 1, 0);
/* Change caps */
GST_VIDEO_INFO_FPS_N (&vinfo) = 50;
GST_VIDEO_INFO_FPS_D (&vinfo) = 1;
caps = gst_video_info_to_caps (&vinfo);
fail_unless (gst_pad_push_event (mysrcpad, gst_event_new_caps (caps)));
gst_caps_unref (caps);
/* Push in data with sink config active, expecting duration calculations
* to be based on the 50 Hz frame rate */
push_data_and_check_output (&sinkcaps_ctx, 192, 192, GST_MSECOND * 80,
GST_MSECOND * 20, 3, 2, 2, 0);
cleanup_rawvideoparse ();
}
GST_END_TEST;
static Suite *
rawvideoparse_suite (void)
{
Suite *s = suite_create ("rawvideoparse");
TCase *tc_chain = tcase_create ("general");
suite_add_tcase (s, tc_chain);
tcase_add_test (tc_chain, test_push_unaligned_data_properties_config);
tcase_add_test (tc_chain, test_push_unaligned_data_sink_caps_config);
tcase_add_test (tc_chain, test_config_switch);
tcase_add_test (tc_chain, test_push_with_no_framerate);
tcase_add_test (tc_chain, test_computed_plane_strides);
tcase_add_test (tc_chain, test_change_caps);
return s;
}
GST_CHECK_MAIN (rawvideoparse);