gstreamer/sys/uvch264/gstuvch264_mjpgdemux.c
2013-02-11 22:05:00 +01:00

717 lines
24 KiB
C

/* GStreamer
*
* uvch264_mjpg_demux: a demuxer for muxed stream in UVC H264 compliant MJPG
*
* Copyright (C) 2012 Cisco Systems, Inc.
* Author: Youness Alaoui <youness.alaoui@collabora.co.uk>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
* Boston, MA 02110-1301, USA.
*/
/**
* SECTION:element-uvch264mjpgdemux
* @short_description: UVC H264 compliant MJPG demuxer
*
* Parses a MJPG stream from a UVC H264 compliant encoding camera and extracts
* each muxed stream into separate pads.
*
*/
#ifdef HAVE_CONFIG_H
#include <config.h>
#endif
#include <string.h>
#include <linux/uvcvideo.h>
#include <linux/usb/video.h>
#include <sys/ioctl.h>
#ifndef UVCIOC_GET_LAST_SCR
#include <time.h>
struct uvc_last_scr_sample
{
__u32 dev_frequency;
__u32 dev_stc;
__u16 dev_sof;
struct timespec host_ts;
__u16 host_sof;
};
#define UVCIOC_GET_LAST_SCR _IOR('u', 0x23, struct uvc_last_scr_sample)
#endif
#include "gstuvch264_mjpgdemux.h"
enum
{
PROP_0,
PROP_DEVICE_FD,
PROP_NUM_CLOCK_SAMPLES
};
#define DEFAULT_NUM_CLOCK_SAMPLES 32
static GstStaticPadTemplate mjpgsink_pad_template =
GST_STATIC_PAD_TEMPLATE ("sink",
GST_PAD_SINK,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("image/jpeg, "
"width = (int) [ 0, MAX ],"
"height = (int) [ 0, MAX ], " "framerate = (fraction) [ 0/1, MAX ] ")
);
static GstStaticPadTemplate jpegsrc_pad_template =
GST_STATIC_PAD_TEMPLATE ("jpeg",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("image/jpeg, "
"width = (int) [ 0, MAX ],"
"height = (int) [ 0, MAX ], " "framerate = (fraction) [ 0/1, MAX ] ")
);
static GstStaticPadTemplate h264src_pad_template =
GST_STATIC_PAD_TEMPLATE ("h264",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("video/x-h264, "
"width = (int) [ 0, MAX ], "
"height = (int) [ 0, MAX ], " "framerate = (fraction) [ 0/1, MAX ] ")
);
static GstStaticPadTemplate yuy2src_pad_template =
GST_STATIC_PAD_TEMPLATE ("yuy2",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("video/x-raw, "
"format = (string) YUY2, "
"width = (int) [ 0, MAX ], "
"height = (int) [ 0, MAX ], " "framerate = (fraction) [ 0/1, MAX ] ")
);
static GstStaticPadTemplate nv12src_pad_template =
GST_STATIC_PAD_TEMPLATE ("nv12",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("video/x-raw, "
"format = (string) NV12, "
"width = (int) [ 0, MAX ], "
"height = (int) [ 0, MAX ], " "framerate = (fraction) [ 0/1, MAX ] ")
);
GST_DEBUG_CATEGORY_STATIC (uvc_h264_mjpg_demux_debug);
#define GST_CAT_DEFAULT uvc_h264_mjpg_demux_debug
typedef struct
{
guint32 dev_stc;
guint32 dev_sof;
GstClockTime host_ts;
guint32 host_sof;
} GstUvcH264ClockSample;
struct _GstUvcH264MjpgDemuxPrivate
{
int device_fd;
int num_clock_samples;
GstUvcH264ClockSample *clock_samples;
int last_sample;
int num_samples;
GstPad *sink_pad;
GstPad *jpeg_pad;
GstPad *h264_pad;
GstPad *yuy2_pad;
GstPad *nv12_pad;
GstCaps *h264_caps;
GstCaps *yuy2_caps;
GstCaps *nv12_caps;
guint16 h264_width;
guint16 h264_height;
guint16 yuy2_width;
guint16 yuy2_height;
guint16 nv12_width;
guint16 nv12_height;
};
typedef struct
{
guint16 version;
guint16 header_len;
guint32 type;
guint16 width;
guint16 height;
guint32 frame_interval;
guint16 delay;
guint32 pts;
} __attribute__ ((packed)) AuxiliaryStreamHeader;
static void gst_uvc_h264_mjpg_demux_set_property (GObject * object,
guint prop_id, const GValue * value, GParamSpec * pspec);
static void gst_uvc_h264_mjpg_demux_get_property (GObject * object,
guint prop_id, GValue * value, GParamSpec * pspec);
static void gst_uvc_h264_mjpg_demux_dispose (GObject * object);
static GstFlowReturn gst_uvc_h264_mjpg_demux_chain (GstPad * pad,
GstObject * parent, GstBuffer * buffer);
static gboolean gst_uvc_h264_mjpg_demux_sink_event (GstPad * pad,
GstObject * parent, GstEvent * event);
static gboolean gst_uvc_h264_mjpg_demux_query (GstPad * pad,
GstObject * parent, GstQuery * query);
#define gst_uvc_h264_mjpg_demux_parent_class parent_class
G_DEFINE_TYPE (GstUvcH264MjpgDemux, gst_uvc_h264_mjpg_demux, GST_TYPE_ELEMENT);
static void
gst_uvc_h264_mjpg_demux_class_init (GstUvcH264MjpgDemuxClass * klass)
{
GObjectClass *gobject_class = (GObjectClass *) klass;
GstElementClass *element_class = (GstElementClass *) klass;
parent_class = g_type_class_peek_parent (klass);
g_type_class_add_private (gobject_class, sizeof (GstUvcH264MjpgDemuxPrivate));
gobject_class->set_property = gst_uvc_h264_mjpg_demux_set_property;
gobject_class->get_property = gst_uvc_h264_mjpg_demux_get_property;
gobject_class->dispose = gst_uvc_h264_mjpg_demux_dispose;
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&mjpgsink_pad_template));
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&jpegsrc_pad_template));
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&h264src_pad_template));
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&yuy2src_pad_template));
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&nv12src_pad_template));
gst_element_class_set_static_metadata (element_class,
"UVC H264 MJPG Demuxer",
"Video/Demuxer",
"Demux UVC H264 auxiliary streams from MJPG images",
"Youness Alaoui <youness.alaoui@collabora.co.uk>");
g_object_class_install_property (gobject_class, PROP_DEVICE_FD,
g_param_spec_int ("device-fd", "device-fd",
"File descriptor of the v4l2 device",
-1, G_MAXINT, -1, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
g_object_class_install_property (gobject_class, PROP_NUM_CLOCK_SAMPLES,
g_param_spec_int ("num-clock-samples", "num-clock-samples",
"Number of clock samples to gather for the PTS synchronization"
" (-1 = unlimited)",
0, G_MAXINT, DEFAULT_NUM_CLOCK_SAMPLES,
G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS));
GST_DEBUG_CATEGORY_INIT (uvc_h264_mjpg_demux_debug,
"uvch264mjpgdemux", 0, "UVC H264 MJPG Demuxer");
}
static void
gst_uvc_h264_mjpg_demux_init (GstUvcH264MjpgDemux * self)
{
self->priv = G_TYPE_INSTANCE_GET_PRIVATE (self, GST_TYPE_UVC_H264_MJPG_DEMUX,
GstUvcH264MjpgDemuxPrivate);
self->priv->device_fd = -1;
/* create the sink and src pads */
self->priv->sink_pad =
gst_pad_new_from_static_template (&mjpgsink_pad_template, "sink");
gst_pad_set_chain_function (self->priv->sink_pad,
GST_DEBUG_FUNCPTR (gst_uvc_h264_mjpg_demux_chain));
gst_pad_set_event_function (self->priv->sink_pad,
GST_DEBUG_FUNCPTR (gst_uvc_h264_mjpg_demux_sink_event));
gst_pad_set_query_function (self->priv->sink_pad,
GST_DEBUG_FUNCPTR (gst_uvc_h264_mjpg_demux_query));
gst_element_add_pad (GST_ELEMENT (self), self->priv->sink_pad);
/* JPEG */
self->priv->jpeg_pad =
gst_pad_new_from_static_template (&jpegsrc_pad_template, "jpeg");
gst_pad_set_query_function (self->priv->jpeg_pad,
GST_DEBUG_FUNCPTR (gst_uvc_h264_mjpg_demux_query));
gst_element_add_pad (GST_ELEMENT (self), self->priv->jpeg_pad);
/* H264 */
self->priv->h264_pad =
gst_pad_new_from_static_template (&h264src_pad_template, "h264");
gst_pad_use_fixed_caps (self->priv->h264_pad);
gst_element_add_pad (GST_ELEMENT (self), self->priv->h264_pad);
/* YUY2 */
self->priv->yuy2_pad =
gst_pad_new_from_static_template (&yuy2src_pad_template, "yuy2");
gst_pad_use_fixed_caps (self->priv->yuy2_pad);
gst_element_add_pad (GST_ELEMENT (self), self->priv->yuy2_pad);
/* NV12 */
self->priv->nv12_pad =
gst_pad_new_from_static_template (&nv12src_pad_template, "nv12");
gst_pad_use_fixed_caps (self->priv->nv12_pad);
gst_element_add_pad (GST_ELEMENT (self), self->priv->nv12_pad);
self->priv->h264_caps = gst_caps_new_empty_simple ("video/x-h264");
self->priv->yuy2_caps = gst_caps_new_simple ("video/x-raw",
"format", G_TYPE_STRING, "YUY2", NULL);
self->priv->nv12_caps = gst_caps_new_simple ("video/x-raw",
"format", G_TYPE_STRING, "NV12", NULL);
self->priv->h264_width = self->priv->h264_height = 0;
self->priv->yuy2_width = self->priv->yuy2_height = 0;
self->priv->nv12_width = self->priv->nv12_height = 0;
}
static void
gst_uvc_h264_mjpg_demux_dispose (GObject * object)
{
GstUvcH264MjpgDemux *self = GST_UVC_H264_MJPG_DEMUX (object);
if (self->priv->h264_caps)
gst_caps_unref (self->priv->h264_caps);
self->priv->h264_caps = NULL;
if (self->priv->yuy2_caps)
gst_caps_unref (self->priv->yuy2_caps);
self->priv->yuy2_caps = NULL;
if (self->priv->nv12_caps)
gst_caps_unref (self->priv->nv12_caps);
self->priv->nv12_caps = NULL;
if (self->priv->clock_samples)
g_free (self->priv->clock_samples);
self->priv->clock_samples = NULL;
G_OBJECT_CLASS (parent_class)->dispose (object);
}
static void
gst_uvc_h264_mjpg_demux_set_property (GObject * object,
guint prop_id, const GValue * value, GParamSpec * pspec)
{
GstUvcH264MjpgDemux *self = GST_UVC_H264_MJPG_DEMUX (object);
switch (prop_id) {
case PROP_DEVICE_FD:
self->priv->device_fd = g_value_get_int (value);
break;
case PROP_NUM_CLOCK_SAMPLES:
self->priv->num_clock_samples = g_value_get_int (value);
if (self->priv->clock_samples) {
if (self->priv->num_clock_samples) {
self->priv->clock_samples = g_realloc_n (self->priv->clock_samples,
self->priv->num_clock_samples, sizeof (GstUvcH264ClockSample));
if (self->priv->num_samples > self->priv->num_clock_samples) {
self->priv->num_samples = self->priv->num_clock_samples;
if (self->priv->last_sample >= self->priv->num_samples)
self->priv->last_sample = self->priv->num_samples - 1;
}
} else {
g_free (self->priv->clock_samples);
self->priv->clock_samples = NULL;
self->priv->last_sample = -1;
self->priv->num_samples = 0;
}
}
if (self->priv->num_clock_samples > 0) {
self->priv->clock_samples = g_malloc0_n (self->priv->num_clock_samples,
sizeof (GstUvcH264ClockSample));
self->priv->last_sample = -1;
self->priv->num_samples = 0;
}
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (self, prop_id, pspec);
break;
}
}
static void
gst_uvc_h264_mjpg_demux_get_property (GObject * object,
guint prop_id, GValue * value, GParamSpec * pspec)
{
GstUvcH264MjpgDemux *self = GST_UVC_H264_MJPG_DEMUX (object);
switch (prop_id) {
case PROP_DEVICE_FD:
g_value_set_int (value, self->priv->device_fd);
break;
case PROP_NUM_CLOCK_SAMPLES:
g_value_set_int (value, self->priv->num_clock_samples);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (self, prop_id, pspec);
break;
}
}
static gboolean
gst_uvc_h264_mjpg_demux_sink_event (GstPad * pad, GstObject * parent,
GstEvent * event)
{
GstUvcH264MjpgDemux *self = GST_UVC_H264_MJPG_DEMUX (parent);
switch (GST_EVENT_TYPE (event)) {
case GST_EVENT_CAPS:
return gst_pad_push_event (self->priv->jpeg_pad, event);
default:
break;
}
return gst_pad_event_default (pad, parent, event);
}
static gboolean
gst_uvc_h264_mjpg_demux_query (GstPad * pad, GstObject * parent,
GstQuery * query)
{
GstUvcH264MjpgDemux *self = GST_UVC_H264_MJPG_DEMUX (parent);
gboolean ret = FALSE;
switch (GST_QUERY_TYPE (query)) {
case GST_QUERY_CAPS:
if (pad == self->priv->sink_pad)
ret = gst_pad_peer_query (self->priv->jpeg_pad, query);
else
ret = gst_pad_peer_query (self->priv->sink_pad, query);
break;
default:
ret = gst_pad_query_default (pad, parent, query);
}
return ret;
}
static gboolean
_pts_to_timestamp (GstUvcH264MjpgDemux * self, GstBuffer * buf, guint32 pts)
{
GstUvcH264MjpgDemuxPrivate *priv = self->priv;
GstUvcH264ClockSample *current_sample = NULL;
GstUvcH264ClockSample *oldest_sample = NULL;
guint32 next_sample;
struct uvc_last_scr_sample sample;
guint32 dev_sof;
if (self->priv->device_fd == -1 || priv->clock_samples == NULL)
return FALSE;
if (-1 == ioctl (priv->device_fd, UVCIOC_GET_LAST_SCR, &sample)) {
//GST_WARNING_OBJECT (self, " GET_LAST_SCR error");
return FALSE;
}
dev_sof = (guint32) (sample.dev_sof + 2048) << 16;
if (priv->num_samples > 0 &&
priv->clock_samples[priv->last_sample].dev_sof == dev_sof) {
current_sample = &priv->clock_samples[priv->last_sample];
} else {
next_sample = (priv->last_sample + 1) % priv->num_clock_samples;
current_sample = &priv->clock_samples[next_sample];
current_sample->dev_stc = sample.dev_stc;
current_sample->dev_sof = dev_sof;
current_sample->host_ts = sample.host_ts.tv_sec * GST_SECOND +
sample.host_ts.tv_nsec * GST_NSECOND;
current_sample->host_sof = (guint32) (sample.host_sof + 2048) << 16;
priv->num_samples++;
priv->last_sample = next_sample;
/* Debug printing */
GST_DEBUG_OBJECT (self, "device frequency: %u", sample.dev_frequency);
GST_DEBUG_OBJECT (self, "dev_sof: %u", sample.dev_sof);
GST_DEBUG_OBJECT (self, "dev_stc: %u", sample.dev_stc);
GST_DEBUG_OBJECT (self, "host_ts: %lu -- %" GST_TIME_FORMAT,
current_sample->host_ts, GST_TIME_ARGS (current_sample->host_ts));
GST_DEBUG_OBJECT (self, "host_sof: %u", sample.host_sof);
GST_DEBUG_OBJECT (self, "PTS: %u", pts);
GST_DEBUG_OBJECT (self, "Diff: %u - %f\n", sample.dev_stc - pts,
(gdouble) (sample.dev_stc - pts) / sample.dev_frequency);
}
if (priv->num_samples < priv->num_clock_samples)
return FALSE;
next_sample = (priv->last_sample + 1) % priv->num_clock_samples;
oldest_sample = &priv->clock_samples[next_sample];
/* TODO: Use current_sample and oldest_sample to do the
* double linear regression and calculate a new PTS */
(void) oldest_sample;
return TRUE;
}
static GstFlowReturn
gst_uvc_h264_mjpg_demux_chain (GstPad * pad,
GstObject * parent, GstBuffer * buf)
{
GstUvcH264MjpgDemux *self;
GstFlowReturn ret = GST_FLOW_OK;
GstBuffer *jpeg_buf = gst_buffer_copy_region (buf, GST_BUFFER_COPY_METADATA,
0, 0);
GstBuffer *aux_buf = NULL;
AuxiliaryStreamHeader aux_header = { 0 };
guint32 aux_size = 0;
GstPad *aux_pad = NULL;
GstCaps **aux_caps = NULL;
guint last_offset;
guint i;
guchar *data;
gsize size;
GstMapInfo info;
self = GST_UVC_H264_MJPG_DEMUX (GST_PAD_PARENT (pad));
last_offset = 0;
size = gst_buffer_get_size (buf);
if (size == 0) {
ret = gst_pad_push (self->priv->jpeg_pad, buf);
goto done;
}
gst_buffer_map (buf, &info, GST_MAP_READ);
data = info.data;
for (i = 0; i < size - 1; i++) {
/* Check for APP4 (0xe4) marker in the jpeg */
if (data[i] == 0xff && data[i + 1] == 0xe4) {
guint16 segment_size;
/* Sanity check sizes and get segment size */
if (i + 4 >= size) {
GST_ELEMENT_ERROR (self, STREAM, DEMUX,
("Not enough data to read marker size"), (NULL));
ret = GST_FLOW_ERROR;
goto done;
}
segment_size = GUINT16_FROM_BE (*((guint16 *) (data + i + 2)));
if (i + segment_size + 2 >= size) {
GST_ELEMENT_ERROR (self, STREAM, DEMUX,
("Not enough data to read marker content"), (NULL));
ret = GST_FLOW_ERROR;
goto done;
}
GST_DEBUG_OBJECT (self,
"Found APP4 marker (%d). JPG: %d-%d - APP4: %d - %d", segment_size,
last_offset, i, i, i + 2 + segment_size);
/* Add JPEG data between the last offset and this market */
if (i - last_offset > 0) {
GstMemory *m = gst_memory_copy (info.memory, last_offset,
i - last_offset);
gst_buffer_append_memory (jpeg_buf, m);
}
last_offset = i + 2 + segment_size;
/* Reset i/segment size to the app4 data (ignore marker header/size) */
i += 4;
segment_size -= 2;
/* If this is a new auxiliary stream, initialize everything properly */
if (aux_buf == NULL) {
if (segment_size < sizeof (aux_header) + sizeof (aux_size)) {
GST_ELEMENT_ERROR (self, STREAM, DEMUX,
("Not enough data to read aux header"), (NULL));
ret = GST_FLOW_ERROR;
goto done;
}
aux_header = *((AuxiliaryStreamHeader *) (data + i));
/* version should be little endian but it looks more like BE */
aux_header.version = GUINT16_FROM_BE (aux_header.version);
aux_header.header_len = GUINT16_FROM_LE (aux_header.header_len);
aux_header.width = GUINT16_FROM_LE (aux_header.width);
aux_header.height = GUINT16_FROM_LE (aux_header.height);
aux_header.frame_interval = GUINT32_FROM_LE (aux_header.frame_interval);
aux_header.delay = GUINT16_FROM_LE (aux_header.delay);
aux_header.pts = GUINT32_FROM_LE (aux_header.pts);
GST_DEBUG_OBJECT (self, "New auxiliary stream : v%d - %d bytes - %"
GST_FOURCC_FORMAT " %dx%d -- %d *100ns -- %d ms -- %d",
aux_header.version, aux_header.header_len,
GST_FOURCC_ARGS (aux_header.type),
aux_header.width, aux_header.height,
aux_header.frame_interval, aux_header.delay, aux_header.pts);
aux_size = *((guint32 *) (data + i + aux_header.header_len));
GST_DEBUG_OBJECT (self, "Auxiliary stream size : %d bytes", aux_size);
if (aux_size > 0) {
guint16 *width = NULL;
guint16 *height = NULL;
/* Find the auxiliary stream's pad and caps */
switch (aux_header.type) {
case GST_MAKE_FOURCC ('H', '2', '6', '4'):
aux_pad = self->priv->h264_pad;
aux_caps = &self->priv->h264_caps;
width = &self->priv->h264_width;
height = &self->priv->h264_height;
break;
case GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'):
aux_pad = self->priv->yuy2_pad;
aux_caps = &self->priv->yuy2_caps;
width = &self->priv->yuy2_width;
height = &self->priv->yuy2_height;
break;
case GST_MAKE_FOURCC ('N', 'V', '1', '2'):
aux_pad = self->priv->nv12_pad;
aux_caps = &self->priv->nv12_caps;
width = &self->priv->nv12_width;
height = &self->priv->nv12_height;
break;
default:
GST_ELEMENT_ERROR (self, STREAM, DEMUX,
("Unknown auxiliary stream format : %" GST_FOURCC_FORMAT,
GST_FOURCC_ARGS (aux_header.type)), (NULL));
ret = GST_FLOW_ERROR;
break;
}
if (ret != GST_FLOW_OK)
goto done;
if (*width != aux_header.width || *height != aux_header.height) {
GstCaps *peercaps = gst_pad_peer_query_caps (aux_pad, NULL);
GstStructure *s = NULL;
gint fps_num = 1000000000 / aux_header.frame_interval;
gint fps_den = 100;
/* TODO: intersect with pad template */
GST_DEBUG ("peercaps : %" GST_PTR_FORMAT, peercaps);
if (peercaps && !gst_caps_is_any (peercaps)) {
peercaps = gst_caps_make_writable (peercaps);
s = gst_caps_get_structure (peercaps, 0);
}
if (s && gst_structure_has_field (s, "framerate")) {
/* TODO: make sure it contains the right format/width/height */
gst_structure_fixate_field_nearest_fraction (s, "framerate",
fps_num, fps_den);
GST_DEBUG ("Fixated struct : %" GST_PTR_FORMAT, s);
gst_structure_get_fraction (s, "framerate", &fps_num, &fps_den);
}
if (peercaps)
gst_caps_unref (peercaps);
*width = aux_header.width;
*height = aux_header.height;
*aux_caps = gst_caps_make_writable (*aux_caps);
/* FIXME: fps must match the caps and be allowed and represent
our first buffer */
gst_caps_set_simple (*aux_caps,
"width", G_TYPE_INT, aux_header.width,
"height", G_TYPE_INT, aux_header.height,
"framerate", GST_TYPE_FRACTION, fps_num, fps_den, NULL);
if (!gst_pad_set_caps (aux_pad, *aux_caps)) {
ret = GST_FLOW_NOT_NEGOTIATED;
goto done;
}
}
/* Create new auxiliary buffer list and adjust i/segment size */
aux_buf = gst_buffer_new ();
}
i += sizeof (aux_header) + sizeof (aux_size);
segment_size -= sizeof (aux_header) + sizeof (aux_size);
}
if (segment_size > aux_size) {
GST_ELEMENT_ERROR (self, STREAM, DEMUX,
("Expected %d auxiliary data, got %d bytes", aux_size,
segment_size), (NULL));
ret = GST_FLOW_ERROR;
goto done;
}
if (segment_size > 0) {
GstMemory *m;
m = gst_memory_copy (info.memory, i, segment_size);
GST_BUFFER_DURATION (aux_buf) =
aux_header.frame_interval * 100 * GST_NSECOND;
_pts_to_timestamp (self, aux_buf, aux_header.pts);
gst_buffer_append_memory (aux_buf, m);
aux_size -= segment_size;
/* Push completed aux data */
if (aux_size == 0) {
GST_DEBUG_OBJECT (self, "Pushing %" GST_FOURCC_FORMAT
" auxiliary buffer %" GST_PTR_FORMAT,
GST_FOURCC_ARGS (aux_header.type), *aux_caps);
ret = gst_pad_push (aux_pad, aux_buf);
aux_buf = NULL;
if (ret != GST_FLOW_OK) {
GST_WARNING_OBJECT (self, "Error pushing %" GST_FOURCC_FORMAT
" auxiliary data", GST_FOURCC_ARGS (aux_header.type));
goto done;
}
}
}
i += segment_size - 1;
} else if (data[i] == 0xff && data[i + 1] == 0xda) {
GstMemory *m;
/* The APP4 markers must be before the SOS marker, so this is the end */
GST_DEBUG_OBJECT (self, "Found SOS marker.");
m = gst_memory_copy (info.memory, last_offset, size - last_offset);
gst_buffer_append_memory (jpeg_buf, m);
last_offset = size;
break;
}
}
if (aux_buf != NULL) {
GST_ELEMENT_ERROR (self, STREAM, DEMUX,
("Incomplete auxiliary stream. %d bytes missing", aux_size), (NULL));
ret = GST_FLOW_ERROR;
goto done;
}
if (last_offset != size) {
/* this means there was no SOS marker in the jpg, so we assume the JPG was
just a container */
GST_DEBUG_OBJECT (self, "SOS marker wasn't found. MJPG is container only");
gst_buffer_unref (jpeg_buf);
jpeg_buf = NULL;
} else {
ret = gst_pad_push (self->priv->jpeg_pad, jpeg_buf);
jpeg_buf = NULL;
}
if (ret != GST_FLOW_OK) {
GST_WARNING_OBJECT (self, "Error pushing jpeg data");
goto done;
}
done:
/* In case of error, unref whatever was left */
if (aux_buf)
gst_buffer_unref (aux_buf);
if (jpeg_buf)
gst_buffer_unref (jpeg_buf);
/* We must always unref the input buffer since we never push it out */
gst_buffer_unref (buf);
return ret;
}