gstreamer/gst/mpeg4videoparse/mpeg4videoparse.c
Sjoerd Simons d0d99f937b gst/mpeg4videoparse/mpeg4videoparse.c: Fix mpeg4videoparse on big endian architectures. Fixes bug #536042.
Original commit message from CVS:
Patch by: Sjoerd Simons <sjoerd at luon dot net>
* gst/mpeg4videoparse/mpeg4videoparse.c:
(gst_mpeg4vparse_handle_vos):
Fix mpeg4videoparse on big endian architectures. Fixes bug #536042.
2008-06-02 07:37:31 +00:00

819 lines
22 KiB
C

/* GStreamer
* Copyright (C) <2008> Mindfruit B.V.
* @author Sjoerd Simons <sjoerd@luon.net>
* Copyright (C) <2007> Julien Moutte <julien@fluendo.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
* Boston, MA 02111-1307, USA.
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#endif
#include <string.h>
#include "mpeg4videoparse.h"
GST_DEBUG_CATEGORY_STATIC (mpeg4v_parse_debug);
#define GST_CAT_DEFAULT mpeg4v_parse_debug
/* elementfactory information */
static GstElementDetails mpeg4vparse_details =
GST_ELEMENT_DETAILS ("MPEG 4 video elementary stream parser",
"Codec/Parser/Video",
"Parses MPEG-4 Part 2 elementary video streams",
"Julien Moutte <julien@fluendo.com>");
static GstStaticPadTemplate src_template =
GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("video/mpeg, "
"mpegversion = (int) 4, "
"parsed = (boolean) true, " "systemstream = (boolean) false")
);
static GstStaticPadTemplate sink_template =
GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("video/mpeg, "
"mpegversion = (int) 4, "
"parsed = (boolean) false, " "systemstream = (boolean) false")
);
/* Properties */
#define DEFAULT_PROP_DROP TRUE
enum
{
PROP_0,
PROP_DROP,
PROP_LAST
};
GST_BOILERPLATE (GstMpeg4VParse, gst_mpeg4vparse, GstElement, GST_TYPE_ELEMENT);
static gboolean
gst_mpeg4vparse_set_new_caps (GstMpeg4VParse * parse,
guint16 time_increment_resolution, guint16 fixed_time_increment,
gint aspect_ratio_width, gint aspect_ratio_height, gint width, gint height)
{
gboolean res;
GstCaps *out_caps = gst_caps_new_simple ("video/mpeg",
"mpegversion", G_TYPE_INT, 4,
"systemstream", G_TYPE_BOOLEAN, FALSE,
"parsed", G_TYPE_BOOLEAN, TRUE, NULL);
if (parse->profile != 0) {
gchar *profile = NULL;
/* FIXME does it make sense to expose the profile in the caps ? */
profile = g_strdup_printf ("%d", parse->profile);
gst_caps_set_simple (out_caps, "profile-level-id",
G_TYPE_STRING, profile, NULL);
g_free (profile);
}
if (parse->config != NULL) {
gst_caps_set_simple (out_caps, "codec_data",
GST_TYPE_BUFFER, parse->config, NULL);
}
if (fixed_time_increment != 0) {
/* we have a framerate */
gst_caps_set_simple (out_caps, "framerate",
GST_TYPE_FRACTION, time_increment_resolution, fixed_time_increment,
NULL);
parse->frame_duration = gst_util_uint64_scale_int (GST_SECOND,
fixed_time_increment, time_increment_resolution);
} else {
/* unknown duration */
parse->frame_duration = 0;
}
if (aspect_ratio_width > 0 && aspect_ratio_height > 0) {
gst_caps_set_simple (out_caps, "pixel-aspect-ratio",
GST_TYPE_FRACTION, aspect_ratio_width, aspect_ratio_height, NULL);
}
if (width > 0 && height > 0) {
gst_caps_set_simple (out_caps,
"width", G_TYPE_INT, width, "height", G_TYPE_INT, height, NULL);
}
GST_DEBUG_OBJECT (parse, "setting downstream caps to %" GST_PTR_FORMAT,
out_caps);
res = gst_pad_set_caps (parse->srcpad, out_caps);
gst_caps_unref (out_caps);
return res;
}
#define VOS_STARTCODE 0xB0
#define VOS_ENDCODE 0xB1
#define USER_DATA_STARTCODE 0xB2
#define GOP_STARTCODE 0xB3
#define VISUAL_OBJECT_STARTCODE 0xB5
#define VOP_STARTCODE 0xB6
#define START_MARKER 0x000001
#define VISUAL_OBJECT_STARTCODE_MARKER ((START_MARKER << 8) + VISUAL_OBJECT_STARTCODE)
typedef struct
{
const guint8 *data;
/* byte offset */
gsize offset;
/* bit offset */
gsize b_offset;
/* size in bytes */
gsize size;
} bitstream_t;
static gboolean
get_bits (bitstream_t * b, int num, guint32 * bits)
{
*bits = 0;
if (b->offset + ((b->b_offset + num) / 8) > b->size)
return FALSE;
if (b->b_offset + num <= 8) {
*bits = b->data[b->offset];
*bits = (*bits >> (8 - num - b->b_offset)) & (((1 << num)) - 1);
b->offset += (b->b_offset + num) / 8;
b->b_offset = (b->b_offset + num) % 8;
return TRUE;
} else {
/* going over the edge.. */
int next;
next = (8 - b->b_offset);
do {
guint32 t;
if (!get_bits (b, next, &t))
return FALSE;
*bits <<= next;
*bits |= t;
num -= next;
next = MIN (8, num);
} while (num > 0);
return TRUE;
}
}
#define GET_BITS(b, num, bits) G_STMT_START { \
if (!get_bits(b, num, bits)) \
goto failed; \
} G_STMT_END
#define MARKER_BIT(b) G_STMT_START { \
guint32 i; \
GET_BITS(b, 1, &i); \
if (i != 0x1) \
goto failed; \
} G_STMT_END
static inline gboolean
next_start_code (bitstream_t * b)
{
guint32 bits;
GET_BITS (b, 1, &bits);
if (bits != 0)
goto failed;
while (b->b_offset != 0) {
GET_BITS (b, 1, &bits);
if (bits != 0x1)
goto failed;
}
return TRUE;
failed:
return FALSE;
}
static gint aspect_ratio_table[6][2] = { {-1, -1}, {1, 1}, {12, 11},
{10, 11}, {16, 11}, {40, 33}
};
/* Returns whether we successfully set the caps downstream if needed */
static gboolean
gst_mpeg4vparse_handle_vos (GstMpeg4VParse * parse, const guint8 * data,
gsize size)
{
/* Skip the startcode */
guint32 bits;
guint16 time_increment_resolution = 0;
guint16 fixed_time_increment = 0;
gint aspect_ratio_width = -1, aspect_ratio_height = -1;
gint height = -1, width = -1;
guint8 profile;
gboolean equal;
bitstream_t bs = { data, 0, 0, size };
/* Parse the config from the VOS frame */
bs.offset = 5;
profile = data[4];
/* invalid profile, yikes */
if (profile == 0)
return FALSE;
equal = FALSE;
if (G_LIKELY (parse->config &&
memcmp (GST_BUFFER_DATA (parse->config), data, size) == 0))
equal = TRUE;
if (G_LIKELY (parse->profile == profile && equal)) {
/* We know this profile and config data, so we can just keep the same caps
*/
return TRUE;
}
/* Even if we fail to parse, then some other element might succeed, so always
* put the VOS in the config */
parse->profile = profile;
if (parse->config != NULL)
gst_buffer_unref (parse->config);
parse->config = gst_buffer_new_and_alloc (size);
memcpy (GST_BUFFER_DATA (parse->config), data, size);
/* Expect Visual Object startcode */
GET_BITS (&bs, 32, &bits);
if (bits != VISUAL_OBJECT_STARTCODE_MARKER)
goto failed;
GET_BITS (&bs, 1, &bits);
if (bits == 0x1) {
/* Skip visual_object_verid and priority */
GET_BITS (&bs, 7, &bits);
}
GET_BITS (&bs, 4, &bits);
/* Only support video ID */
if (bits != 0x1)
goto failed;
/* video signal type */
GET_BITS (&bs, 1, &bits);
if (bits == 0x1) {
/* video signal type, ignore format and range */
GET_BITS (&bs, 4, &bits);
GET_BITS (&bs, 1, &bits);
if (bits == 0x1) {
/* ignore color description */
GET_BITS (&bs, 24, &bits);
}
}
if (!next_start_code (&bs))
goto failed;
/* expecting a video object startcode */
GET_BITS (&bs, 32, &bits);
if (bits > 0x11F)
goto failed;
/* expecting a video object layer startcode */
GET_BITS (&bs, 32, &bits);
if (bits < 0x120 || bits > 0x12F)
goto failed;
/* ignore random accessible vol and video object type indication */
GET_BITS (&bs, 9, &bits);
GET_BITS (&bs, 1, &bits);
if (bits) {
/* skip video object layer verid and priority */
GET_BITS (&bs, 7, &bits);
}
/* aspect ratio info */
GET_BITS (&bs, 4, &bits);
if (bits == 0)
goto failed;
/* check if aspect ratio info is extended par */
if (bits == 0xff) {
GET_BITS (&bs, 4, &bits);
aspect_ratio_width = bits;
GET_BITS (&bs, 4, &bits);
aspect_ratio_height = bits;
} else if (bits < 0x6) {
aspect_ratio_width = aspect_ratio_table[bits][0];
aspect_ratio_height = aspect_ratio_table[bits][1];
}
GET_BITS (&bs, 1, &bits);
if (bits) {
/* vol control parameters, skip chroma and low delay */
GET_BITS (&bs, 3, &bits);
GET_BITS (&bs, 1, &bits);
if (bits) {
/* skip vbv_parameters */
GET_BITS (&bs, 79, &bits);
}
}
/* layer shape */
GET_BITS (&bs, 2, &bits);
/* only support rectangular */
if (bits != 0)
goto failed;
MARKER_BIT (&bs);
GET_BITS (&bs, 16, &bits);
time_increment_resolution = bits;
MARKER_BIT (&bs);
GST_DEBUG_OBJECT (parse, "time increment resolution %d",
time_increment_resolution);
GET_BITS (&bs, 1, &bits);
if (bits) {
/* fixed time increment */
int n;
/* Lenght of the time increment is the minimal number of bits needed to
* represent time_increment_resolution */
for (n = 0; (time_increment_resolution >> n) != 0; n++);
GET_BITS (&bs, n, &bits);
fixed_time_increment = bits;
} else {
fixed_time_increment = 1;
}
GST_DEBUG_OBJECT (parse, "fixed time increment %d", fixed_time_increment);
/* assuming rectangular shape */
MARKER_BIT (&bs);
GET_BITS (&bs, 13, &bits);
width = bits;
MARKER_BIT (&bs);
GET_BITS (&bs, 13, &bits);
height = bits;
MARKER_BIT (&bs);
out:
return gst_mpeg4vparse_set_new_caps (parse, time_increment_resolution,
fixed_time_increment, aspect_ratio_width, aspect_ratio_height,
width, height);
failed:
GST_WARNING_OBJECT (parse, "Failed to parse config data");
goto out;
}
static void
gst_mpeg4vparse_push (GstMpeg4VParse * parse, gsize size)
{
if (G_UNLIKELY (parse->config == NULL && parse->drop)) {
GST_LOG_OBJECT (parse, "Dropping %d bytes", parse->offset);
gst_adapter_flush (parse->adapter, size);
} else {
GstBuffer *out_buf;
out_buf = gst_adapter_take_buffer (parse->adapter, parse->offset);
if (out_buf) {
/* Set GST_BUFFER_FLAG_DELTA_UNIT if it's not an intra frame */
if (!parse->intra_frame) {
GST_BUFFER_FLAG_SET (out_buf, GST_BUFFER_FLAG_DELTA_UNIT);
}
gst_buffer_set_caps (out_buf, GST_PAD_CAPS (parse->srcpad));
GST_BUFFER_TIMESTAMP (out_buf) = parse->timestamp;
gst_pad_push (parse->srcpad, out_buf);
}
}
/* Restart now that we flushed data */
parse->offset = 0;
parse->state = PARSE_NEED_START;
parse->intra_frame = FALSE;
}
static GstFlowReturn
gst_mpeg4vparse_drain (GstMpeg4VParse * parse, GstBuffer * last_buffer)
{
GstFlowReturn ret = GST_FLOW_OK;
const guint8 *data = NULL;
guint available = 0;
available = gst_adapter_available (parse->adapter);
/* We do a quick check here to avoid the _peek() below. */
if (G_UNLIKELY (available < 5)) {
GST_DEBUG_OBJECT (parse, "we need more data, %d < 5", available);
goto beach;
}
data = gst_adapter_peek (parse->adapter, available);
/* Need at least 5 more bytes, 4 for the startcode, 1 to optionally determine
* the VOP frame type */
while (available >= 5 && parse->offset < available - 5) {
if (data[parse->offset] == 0 && data[parse->offset + 1] == 0 &&
data[parse->offset + 2] == 1) {
switch (parse->state) {
case PARSE_NEED_START:
switch (data[parse->offset + 3]) {
case VOP_STARTCODE:
case VOS_STARTCODE:
case GOP_STARTCODE:
/* valid starts of a frame */
parse->state = PARSE_START_FOUND;
if (parse->offset > 0) {
GST_LOG_OBJECT (parse, "Flushing %u bytes", parse->offset);
gst_adapter_flush (parse->adapter, parse->offset);
parse->offset = 0;
available = gst_adapter_available (parse->adapter);
data = gst_adapter_peek (parse->adapter, available);
}
break;
default:
parse->offset += 4;
}
break;
case PARSE_START_FOUND:
switch (data[parse->offset + 3]) {
case VOP_STARTCODE:
GST_LOG_OBJECT (parse, "found VOP start marker at %u",
parse->offset);
parse->intra_frame = ((data[parse->offset + 4] >> 6 & 0x3) == 0);
/* Ensure that the timestamp of the outgoing buffer is the same
* as the one the VOP header is found in */
parse->timestamp = GST_BUFFER_TIMESTAMP (last_buffer);
parse->state = PARSE_VOP_FOUND;
break;
case VOS_STARTCODE:
GST_LOG_OBJECT (parse, "found VOS start marker at %u",
parse->offset);
parse->vos_offset = parse->offset;
parse->state = PARSE_VOS_FOUND;
break;
}
/* Jump over it */
parse->offset += 4;
break;
case PARSE_VOS_FOUND:
switch (data[parse->offset + 3]) {
case GOP_STARTCODE:
case VOP_STARTCODE:
/* end of VOS found, interpret the config data and restart the
* search for the VOP */
gst_mpeg4vparse_handle_vos (parse, data + parse->vos_offset,
parse->offset - parse->vos_offset);
parse->state = PARSE_START_FOUND;
break;
default:
parse->offset += 4;
}
break;
case PARSE_VOP_FOUND:
{ /* We were in a VOP already, any start code marks the end of it */
GST_LOG_OBJECT (parse, "found VOP end marker at %u", parse->offset);
gst_mpeg4vparse_push (parse, parse->offset);
available = gst_adapter_available (parse->adapter);
data = gst_adapter_peek (parse->adapter, available);
break;
}
default:
GST_WARNING_OBJECT (parse, "unexpected parse state (%d)",
parse->state);
ret = GST_FLOW_UNEXPECTED;
goto beach;
}
} else { /* Continue searching */
parse->offset++;
}
}
beach:
return ret;
}
static GstFlowReturn
gst_mpeg4vparse_chain (GstPad * pad, GstBuffer * buffer)
{
GstMpeg4VParse *parse = GST_MPEG4VIDEOPARSE (gst_pad_get_parent (pad));
GstFlowReturn ret = GST_FLOW_OK;
GST_DEBUG_OBJECT (parse, "received buffer of %u bytes with ts %"
GST_TIME_FORMAT " and offset %" G_GINT64_FORMAT, GST_BUFFER_SIZE (buffer),
GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (buffer)),
GST_BUFFER_OFFSET (buffer));
gst_adapter_push (parse->adapter, buffer);
/* Drain the accumulated blocks frame per frame */
ret = gst_mpeg4vparse_drain (parse, buffer);
gst_object_unref (parse);
return ret;
}
static gboolean
gst_mpeg4vparse_sink_setcaps (GstPad * pad, GstCaps * caps)
{
gboolean res = TRUE;
GstMpeg4VParse *parse = GST_MPEG4VIDEOPARSE (gst_pad_get_parent (pad));
GstStructure *s;
const GValue *value;
GST_DEBUG_OBJECT (parse, "setcaps called with %" GST_PTR_FORMAT, caps);
s = gst_caps_get_structure (caps, 0);
if ((value = gst_structure_get_value (s, "codec_data")) != NULL
&& G_VALUE_HOLDS (value, GST_TYPE_BUFFER)) {
GstBuffer *buf = gst_value_get_buffer (value);
res = gst_mpeg4vparse_handle_vos (parse, GST_BUFFER_DATA (buf),
GST_BUFFER_SIZE (buf));
} else {
/* No codec data, set minimal new caps.. VOS parsing later will fill in
* the other fields */
res = gst_mpeg4vparse_set_new_caps (parse, 0, 0, 0, 0, 0, 0);
}
gst_object_unref (parse);
return res;
}
static gboolean
gst_mpeg4vparse_sink_event (GstPad * pad, GstEvent * event)
{
gboolean res = TRUE;
GstMpeg4VParse *parse = GST_MPEG4VIDEOPARSE (gst_pad_get_parent (pad));
GST_DEBUG_OBJECT (parse, "handling event type %s",
GST_EVENT_TYPE_NAME (event));
switch (GST_EVENT_TYPE (event)) {
case GST_EVENT_EOS:
if (parse->state == PARSE_VOP_FOUND) {
/* If we've found the start of the VOP assume what's left in the
* adapter is the complete VOP. This might cause us to send an
* incomplete VOP out, but prevents the last video frame from
* potentially being dropped */
gst_mpeg4vparse_push (parse, gst_adapter_available (parse->adapter));
}
/* fallthrough */
default:
res = gst_pad_event_default (pad, event);
break;
}
gst_object_unref (parse);
return res;
}
static gboolean
gst_mpeg4vparse_src_query (GstPad * pad, GstQuery * query)
{
GstMpeg4VParse *parse = GST_MPEG4VIDEOPARSE (gst_pad_get_parent (pad));
gboolean res;
switch (GST_QUERY_TYPE (query)) {
case GST_QUERY_LATENCY:
{
/* We need to send the query upstream and add the returned latency to our
* own */
GstClockTime min_latency, max_latency;
gboolean us_live;
GstClockTime our_latency;
if ((res = gst_pad_peer_query (parse->sinkpad, query))) {
gst_query_parse_latency (query, &us_live, &min_latency, &max_latency);
GST_DEBUG_OBJECT (parse, "Peer latency: min %"
GST_TIME_FORMAT " max %" GST_TIME_FORMAT,
GST_TIME_ARGS (min_latency), GST_TIME_ARGS (max_latency));
/* our latency is 1 frame, find the frame duration */
our_latency = parse->frame_duration;
GST_DEBUG_OBJECT (parse, "Our latency: %" GST_TIME_FORMAT,
GST_TIME_ARGS (our_latency));
/* we add some latency */
min_latency += our_latency;
if (max_latency != -1)
max_latency += our_latency;
GST_DEBUG_OBJECT (parse, "Calculated total latency : min %"
GST_TIME_FORMAT " max %" GST_TIME_FORMAT,
GST_TIME_ARGS (min_latency), GST_TIME_ARGS (max_latency));
gst_query_set_latency (query, us_live, min_latency, max_latency);
}
break;
}
default:
res = gst_pad_peer_query (parse->sinkpad, query);
break;
}
gst_object_unref (parse);
return res;
}
static void
gst_mpeg4vparse_cleanup (GstMpeg4VParse * parse)
{
if (parse->adapter) {
gst_adapter_clear (parse->adapter);
}
parse->state = PARSE_NEED_START;
parse->offset = 0;
}
static GstStateChangeReturn
gst_mpeg4vparse_change_state (GstElement * element, GstStateChange transition)
{
GstMpeg4VParse *parse = GST_MPEG4VIDEOPARSE (element);
GstStateChangeReturn ret;
ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
switch (transition) {
case GST_STATE_CHANGE_PAUSED_TO_READY:
gst_mpeg4vparse_cleanup (parse);
break;
default:
break;
}
return ret;
}
static void
gst_mpeg4vparse_dispose (GObject * object)
{
GstMpeg4VParse *parse = GST_MPEG4VIDEOPARSE (object);
if (parse->adapter) {
g_object_unref (parse->adapter);
parse->adapter = NULL;
}
GST_CALL_PARENT (G_OBJECT_CLASS, dispose, (object));
}
static void
gst_mpeg4vparse_base_init (gpointer klass)
{
GstElementClass *element_class = GST_ELEMENT_CLASS (klass);
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&src_template));
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&sink_template));
gst_element_class_set_details (element_class, &mpeg4vparse_details);
}
static void
gst_mpeg4vparse_set_property (GObject * object, guint property_id,
const GValue * value, GParamSpec * pspec)
{
GstMpeg4VParse *parse = GST_MPEG4VIDEOPARSE (object);
switch (property_id) {
case PROP_DROP:
parse->drop = g_value_get_boolean (value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec);
}
}
static void
gst_mpeg4vparse_get_property (GObject * object, guint property_id,
GValue * value, GParamSpec * pspec)
{
GstMpeg4VParse *parse = GST_MPEG4VIDEOPARSE (object);
switch (property_id) {
case PROP_DROP:
g_value_set_boolean (value, parse->drop);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec);
}
}
static void
gst_mpeg4vparse_class_init (GstMpeg4VParseClass * klass)
{
GObjectClass *gobject_class;
GstElementClass *gstelement_class;
gstelement_class = (GstElementClass *) klass;
gobject_class = G_OBJECT_CLASS (klass);
parent_class = g_type_class_peek_parent (klass);
gobject_class->dispose = GST_DEBUG_FUNCPTR (gst_mpeg4vparse_dispose);
gobject_class->set_property = gst_mpeg4vparse_set_property;
gobject_class->get_property = gst_mpeg4vparse_get_property;
g_object_class_install_property (gobject_class, PROP_DROP,
g_param_spec_boolean ("drop", "drop",
"Drop data untill valid configuration data is received either "
"in the stream or through caps", DEFAULT_PROP_DROP,
G_PARAM_CONSTRUCT | G_PARAM_READWRITE));
gstelement_class->change_state =
GST_DEBUG_FUNCPTR (gst_mpeg4vparse_change_state);
}
static void
gst_mpeg4vparse_init (GstMpeg4VParse * parse, GstMpeg4VParseClass * g_class)
{
parse->sinkpad = gst_pad_new_from_static_template (&sink_template, "sink");
gst_pad_set_chain_function (parse->sinkpad,
GST_DEBUG_FUNCPTR (gst_mpeg4vparse_chain));
gst_pad_set_event_function (parse->sinkpad,
GST_DEBUG_FUNCPTR (gst_mpeg4vparse_sink_event));
gst_pad_set_setcaps_function (parse->sinkpad,
GST_DEBUG_FUNCPTR (gst_mpeg4vparse_sink_setcaps));
gst_element_add_pad (GST_ELEMENT (parse), parse->sinkpad);
parse->srcpad = gst_pad_new_from_static_template (&src_template, "src");
gst_pad_set_query_function (parse->srcpad,
GST_DEBUG_FUNCPTR (gst_mpeg4vparse_src_query));
gst_pad_use_fixed_caps (parse->srcpad);
gst_element_add_pad (GST_ELEMENT (parse), parse->srcpad);
parse->adapter = gst_adapter_new ();
gst_mpeg4vparse_cleanup (parse);
}
static gboolean
plugin_init (GstPlugin * plugin)
{
GST_DEBUG_CATEGORY_INIT (mpeg4v_parse_debug, "mpeg4videoparse", 0,
"MPEG-4 video parser");
if (!gst_element_register (plugin, "mpeg4videoparse", GST_RANK_SECONDARY,
gst_mpeg4vparse_get_type ()))
return FALSE;
return TRUE;
}
GST_PLUGIN_DEFINE (GST_VERSION_MAJOR,
GST_VERSION_MINOR,
"mpeg4videoparse",
"MPEG-4 video parser",
plugin_init, VERSION, GST_LICENSE, GST_PACKAGE_NAME, GST_PACKAGE_ORIGIN)