gstreamer/gst/avi/gstavimux.c
Thiago Santos b59dc3e5fb avimux: also set the suggested buf size for audio
We were only setting the suggested buf size for video,
we can set it for audio as well.

This and 195e14529d80ef318ce3a778c1995efb11f266cd
fix an issue that prevented seeking on large avi files
on WMP (non-recent versions).
2009-11-24 12:44:57 -03:00

2149 lines
68 KiB
C

/* AVI muxer plugin for GStreamer
* Copyright (C) 2002 Ronald Bultje <rbultje@ronald.bitfreak.net>
* (C) 2006 Mark Nauwelaerts <manauw@skynet.be>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
* Boston, MA 02111-1307, USA.
*/
/* based on:
* - the old avimuxer (by Wim Taymans)
* - xawtv's aviwriter (by Gerd Knorr)
* - mjpegtools' avilib (by Rainer Johanni)
* - openDML large-AVI docs
*/
/**
* SECTION:element-avimux
*
* Muxes raw or compressed audio and/or video streams into an AVI file.
*
* <refsect2>
* <title>Example launch lines</title>
* <para>(write everything in one line, without the backslash characters)</para>
* |[
* gst-launch videotestsrc num-buffers=250 \
* ! 'video/x-raw-yuv,format=(fourcc)I420,width=320,height=240,framerate=(fraction)25/1' \
* ! queue ! mux. \
* audiotestsrc num-buffers=440 ! audioconvert \
* ! 'audio/x-raw-int,rate=44100,channels=2' ! queue ! mux. \
* avimux name=mux ! filesink location=test.avi
* ]| This will create an .AVI file containing an uncompressed video stream
* with a test picture and an uncompressed audio stream containing a
* test sound.
* |[
* gst-launch videotestsrc num-buffers=250 \
* ! 'video/x-raw-yuv,format=(fourcc)I420,width=320,height=240,framerate=(fraction)25/1' \
* ! xvidenc ! queue ! mux. \
* audiotestsrc num-buffers=440 ! audioconvert ! 'audio/x-raw-int,rate=44100,channels=2' \
* ! lame ! queue ! mux. \
* avimux name=mux ! filesink location=test.avi
* ]| This will create an .AVI file containing the same test video and sound
* as above, only that both streams will be compressed this time. This will
* only work if you have the necessary encoder elements installed of course.
* </refsect2>
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#endif
#include "gst/gst-i18n-plugin.h"
#include <stdlib.h>
#include <string.h>
#include <gst/video/video.h>
#include "gstavimux.h"
GST_DEBUG_CATEGORY_STATIC (avimux_debug);
#define GST_CAT_DEFAULT avimux_debug
enum
{
ARG_0,
ARG_BIGFILE
};
#define DEFAULT_BIGFILE TRUE
static const GstElementDetails gst_avi_mux_details =
GST_ELEMENT_DETAILS ("Avi muxer",
"Codec/Muxer",
"Muxes audio and video into an avi stream",
"GStreamer maintainers <gstreamer-devel@lists.sourceforge.net>");
static GstStaticPadTemplate src_factory = GST_STATIC_PAD_TEMPLATE ("src",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("video/x-msvideo")
);
static GstStaticPadTemplate video_sink_factory =
GST_STATIC_PAD_TEMPLATE ("video_%d",
GST_PAD_SINK,
GST_PAD_REQUEST,
GST_STATIC_CAPS ("video/x-raw-yuv, "
"format = (fourcc) { YUY2, I420 }, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ]; "
"image/jpeg, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ]; "
"video/x-divx, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ], "
"divxversion = (int) [ 3, 5 ]; "
"video/x-xvid, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ]; "
"video/x-3ivx, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ]; "
"video/x-msmpeg, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ], "
"msmpegversion = (int) [ 41, 43 ]; "
"video/mpeg, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ], "
"mpegversion = (int) { 1, 2, 4}, "
"systemstream = (boolean) FALSE; "
"video/x-h263, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ]; "
"video/x-h264, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], "
"framerate = (fraction) [ 0, MAX ]; "
"video/x-dv, "
"width = (int) 720, "
"height = (int) { 576, 480 }, "
"framerate = (fraction) [ 0, MAX ], "
"systemstream = (boolean) FALSE; "
"video/x-huffyuv, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], " "framerate = (fraction) [ 0, MAX ];"
"video/x-dirac, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], " "framerate = (fraction) [ 0, MAX ];"
"video/x-wmv, "
"width = (int) [ 16, 4096 ], "
"height = (int) [ 16, 4096 ], " "framerate = (fraction) [ 0, MAX ], "
"wmvversion = (int) [ 1, 3];"
"image/x-jpc, "
"width = (int) [ 1, 2147483647 ], "
"height = (int) [ 1, 2147483647 ], "
"framerate = (fraction) [ 0, MAX ]")
);
static GstStaticPadTemplate audio_sink_factory =
GST_STATIC_PAD_TEMPLATE ("audio_%d",
GST_PAD_SINK,
GST_PAD_REQUEST,
GST_STATIC_CAPS ("audio/x-raw-int, "
"endianness = (int) LITTLE_ENDIAN, "
"signed = (boolean) { TRUE, FALSE }, "
"width = (int) { 8, 16 }, "
"depth = (int) { 8, 16 }, "
"rate = (int) [ 1000, 96000 ], "
"channels = (int) [ 1, 2 ]; "
"audio/mpeg, "
"mpegversion = (int) 1, "
"layer = (int) [ 1, 3 ], "
"rate = (int) [ 1000, 96000 ], " "channels = (int) [ 1, 2 ]; "
"audio/mpeg, "
"mpegversion = (int) 4, "
"rate = (int) [ 1000, 96000 ], " "channels = (int) [ 1, 2 ]; "
/*#if 0 VC6 doesn't support #if here ...
"audio/x-vorbis, "
"rate = (int) [ 1000, 96000 ], " "channels = (int) [ 1, 2 ]; "
#endif*/
"audio/x-ac3, "
"rate = (int) [ 1000, 96000 ], " "channels = (int) [ 1, 2 ]; "
"audio/x-alaw, "
"rate = (int) [ 1000, 48000 ], " "channels = (int) [ 1, 2 ]; "
"audio/x-mulaw, "
"rate = (int) [ 1000, 48000 ], " "channels = (int) [ 1, 2 ]; "
"audio/x-wma, "
"rate = (int) [ 1000, 96000 ], " "channels = (int) [ 1, 2 ], "
"wmaversion = (int) [ 1, 2 ] ")
);
static void gst_avi_mux_base_init (gpointer g_class);
static void gst_avi_mux_class_init (GstAviMuxClass * klass);
static void gst_avi_mux_init (GstAviMux * avimux);
static void gst_avi_mux_pad_reset (GstAviPad * avipad, gboolean free);
static GstFlowReturn gst_avi_mux_collect_pads (GstCollectPads * pads,
GstAviMux * avimux);
static gboolean gst_avi_mux_handle_event (GstPad * pad, GstEvent * event);
static GstPad *gst_avi_mux_request_new_pad (GstElement * element,
GstPadTemplate * templ, const gchar * name);
static void gst_avi_mux_release_pad (GstElement * element, GstPad * pad);
static void gst_avi_mux_set_property (GObject * object,
guint prop_id, const GValue * value, GParamSpec * pspec);
static void gst_avi_mux_get_property (GObject * object,
guint prop_id, GValue * value, GParamSpec * pspec);
static GstStateChangeReturn gst_avi_mux_change_state (GstElement * element,
GstStateChange transition);
static GstElementClass *parent_class = NULL;
GType
gst_avi_mux_get_type (void)
{
static GType avimux_type = 0;
if (!avimux_type) {
static const GTypeInfo avimux_info = {
sizeof (GstAviMuxClass),
gst_avi_mux_base_init,
NULL,
(GClassInitFunc) gst_avi_mux_class_init,
NULL,
NULL,
sizeof (GstAviMux),
0,
(GInstanceInitFunc) gst_avi_mux_init,
};
static const GInterfaceInfo tag_setter_info = {
NULL,
NULL,
NULL
};
avimux_type =
g_type_register_static (GST_TYPE_ELEMENT, "GstAviMux", &avimux_info, 0);
g_type_add_interface_static (avimux_type, GST_TYPE_TAG_SETTER,
&tag_setter_info);
}
return avimux_type;
}
static void
gst_avi_mux_base_init (gpointer g_class)
{
GstElementClass *element_class = GST_ELEMENT_CLASS (g_class);
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&src_factory));
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&audio_sink_factory));
gst_element_class_add_pad_template (element_class,
gst_static_pad_template_get (&video_sink_factory));
gst_element_class_set_details (element_class, &gst_avi_mux_details);
GST_DEBUG_CATEGORY_INIT (avimux_debug, "avimux", 0, "Muxer for AVI streams");
}
static void
gst_avi_mux_finalize (GObject * object)
{
GstAviMux *mux = GST_AVI_MUX (object);
GSList *node;
/* completely free each sinkpad */
node = mux->sinkpads;
while (node) {
GstAviPad *avipad = (GstAviPad *) node->data;
node = node->next;
gst_avi_mux_pad_reset (avipad, TRUE);
g_free (avipad);
}
g_slist_free (mux->sinkpads);
mux->sinkpads = NULL;
g_free (mux->idx);
mux->idx = NULL;
gst_object_unref (mux->collect);
G_OBJECT_CLASS (parent_class)->finalize (object);
}
static void
gst_avi_mux_class_init (GstAviMuxClass * klass)
{
GObjectClass *gobject_class;
GstElementClass *gstelement_class;
gobject_class = (GObjectClass *) klass;
gstelement_class = (GstElementClass *) klass;
parent_class = g_type_class_peek_parent (klass);
gobject_class->get_property = gst_avi_mux_get_property;
gobject_class->set_property = gst_avi_mux_set_property;
gobject_class->finalize = gst_avi_mux_finalize;
g_object_class_install_property (gobject_class, ARG_BIGFILE,
g_param_spec_boolean ("bigfile", "Bigfile Support (>2GB)",
"Support for openDML-2.0 (big) AVI files", DEFAULT_BIGFILE,
G_PARAM_READWRITE));
gstelement_class->request_new_pad =
GST_DEBUG_FUNCPTR (gst_avi_mux_request_new_pad);
gstelement_class->release_pad = GST_DEBUG_FUNCPTR (gst_avi_mux_release_pad);
gstelement_class->change_state = GST_DEBUG_FUNCPTR (gst_avi_mux_change_state);
}
/* reset pad to initial state
* free - if true, release all, not only stream related, data */
static void
gst_avi_mux_pad_reset (GstAviPad * avipad, gboolean free)
{
/* generic part */
memset (&(avipad->hdr), 0, sizeof (gst_riff_strh));
memset (&(avipad->idx[0]), 0, sizeof (avipad->idx));
if (free) {
g_free (avipad->tag);
avipad->tag = NULL;
g_free (avipad->idx_tag);
avipad->idx_tag = NULL;
}
if (avipad->is_video) {
GstAviVideoPad *vidpad = (GstAviVideoPad *) avipad;
avipad->hdr.type = GST_MAKE_FOURCC ('v', 'i', 'd', 's');
if (vidpad->vids_codec_data) {
gst_buffer_unref (vidpad->vids_codec_data);
vidpad->vids_codec_data = NULL;
}
memset (&(vidpad->vids), 0, sizeof (gst_riff_strf_vids));
memset (&(vidpad->vprp), 0, sizeof (gst_riff_vprp));
} else {
GstAviAudioPad *audpad = (GstAviAudioPad *) avipad;
audpad->samples = 0;
avipad->hdr.type = GST_MAKE_FOURCC ('a', 'u', 'd', 's');
if (audpad->auds_codec_data) {
gst_buffer_unref (audpad->auds_codec_data);
audpad->auds_codec_data = NULL;
}
memset (&(audpad->auds), 0, sizeof (gst_riff_strf_auds));
}
}
static void
gst_avi_mux_reset (GstAviMux * avimux)
{
GSList *node, *newlist = NULL;
/* free and reset each sinkpad */
node = avimux->sinkpads;
while (node) {
GstAviPad *avipad = (GstAviPad *) node->data;
node = node->next;
gst_avi_mux_pad_reset (avipad, FALSE);
/* if this pad has collectdata, keep it, otherwise dump it completely */
if (avipad->collect)
newlist = g_slist_append (newlist, avipad);
else {
gst_avi_mux_pad_reset (avipad, TRUE);
g_free (avipad);
}
}
/* free the old list of sinkpads, only keep the real collecting ones */
g_slist_free (avimux->sinkpads);
avimux->sinkpads = newlist;
/* avi data */
avimux->num_frames = 0;
memset (&(avimux->avi_hdr), 0, sizeof (gst_riff_avih));
avimux->avi_hdr.max_bps = 10000000;
avimux->codec_data_size = 0;
if (avimux->tags_snap) {
gst_tag_list_free (avimux->tags_snap);
avimux->tags_snap = NULL;
}
g_free (avimux->idx);
avimux->idx = NULL;
/* state info */
avimux->write_header = TRUE;
/* tags */
gst_tag_setter_reset_tags (GST_TAG_SETTER (avimux));
}
static void
gst_avi_mux_init (GstAviMux * avimux)
{
avimux->srcpad = gst_pad_new_from_static_template (&src_factory, "src");
gst_pad_use_fixed_caps (avimux->srcpad);
gst_element_add_pad (GST_ELEMENT (avimux), avimux->srcpad);
/* property */
avimux->enable_large_avi = DEFAULT_BIGFILE;
avimux->collect = gst_collect_pads_new ();
gst_collect_pads_set_function (avimux->collect,
(GstCollectPadsFunction) (GST_DEBUG_FUNCPTR (gst_avi_mux_collect_pads)),
avimux);
/* set to clean state */
gst_avi_mux_reset (avimux);
}
static gboolean
gst_avi_mux_vidsink_set_caps (GstPad * pad, GstCaps * vscaps)
{
GstAviMux *avimux;
GstAviVideoPad *avipad;
GstAviCollectData *collect_pad;
GstStructure *structure;
const gchar *mimetype;
const GValue *fps, *par;
const GValue *codec_data;
gint width, height;
gint par_n, par_d;
avimux = GST_AVI_MUX (gst_pad_get_parent (pad));
/* find stream data */
collect_pad = (GstAviCollectData *) gst_pad_get_element_private (pad);
g_assert (collect_pad);
avipad = (GstAviVideoPad *) collect_pad->avipad;
g_assert (avipad);
g_assert (avipad->parent.is_video);
g_assert (avipad->parent.hdr.type == GST_MAKE_FOURCC ('v', 'i', 'd', 's'));
GST_DEBUG_OBJECT (avimux, "%s:%s, caps=%" GST_PTR_FORMAT,
GST_DEBUG_PAD_NAME (pad), vscaps);
structure = gst_caps_get_structure (vscaps, 0);
mimetype = gst_structure_get_name (structure);
/* global */
avipad->vids.size = sizeof (gst_riff_strf_vids);
avipad->vids.planes = 1;
if (!gst_structure_get_int (structure, "width", &width) ||
!gst_structure_get_int (structure, "height", &height)) {
goto refuse_caps;
}
avipad->vids.width = width;
avipad->vids.height = height;
fps = gst_structure_get_value (structure, "framerate");
if (fps == NULL || !GST_VALUE_HOLDS_FRACTION (fps))
goto refuse_caps;
avipad->parent.hdr.rate = gst_value_get_fraction_numerator (fps);
avipad->parent.hdr.scale = gst_value_get_fraction_denominator (fps);
/* (pixel) aspect ratio data, if any */
par = gst_structure_get_value (structure, "pixel-aspect-ratio");
/* only use video properties header if there is non-trivial aspect info */
if (par && GST_VALUE_HOLDS_FRACTION (par) &&
((par_n = gst_value_get_fraction_numerator (par)) !=
(par_d = gst_value_get_fraction_denominator (par)))) {
GValue to_ratio = { 0, };
guint ratio_n, ratio_d;
/* some fraction voodoo to obtain simplest possible ratio */
g_value_init (&to_ratio, GST_TYPE_FRACTION);
gst_value_set_fraction (&to_ratio, width * par_n, height * par_d);
ratio_n = gst_value_get_fraction_numerator (&to_ratio);
ratio_d = gst_value_get_fraction_denominator (&to_ratio);
GST_DEBUG_OBJECT (avimux, "generating vprp data with aspect ratio %d/%d",
ratio_n, ratio_d);
/* simply fill in */
avipad->vprp.vert_rate = avipad->parent.hdr.rate / avipad->parent.hdr.scale;
avipad->vprp.hor_t_total = width;
avipad->vprp.vert_lines = height;
avipad->vprp.aspect = (ratio_n) << 16 | (ratio_d & 0xffff);
avipad->vprp.width = width;
avipad->vprp.height = height;
avipad->vprp.fields = 1;
avipad->vprp.field_info[0].compressed_bm_height = height;
avipad->vprp.field_info[0].compressed_bm_width = width;
avipad->vprp.field_info[0].valid_bm_height = height;
avipad->vprp.field_info[0].valid_bm_width = width;
}
/* codec initialization data, if any */
codec_data = gst_structure_get_value (structure, "codec_data");
if (codec_data) {
avipad->vids_codec_data = gst_value_get_buffer (codec_data);
gst_buffer_ref (avipad->vids_codec_data);
/* keep global track of size */
avimux->codec_data_size += GST_BUFFER_SIZE (avipad->vids_codec_data);
}
if (!strcmp (mimetype, "video/x-raw-yuv")) {
guint32 format;
gst_structure_get_fourcc (structure, "format", &format);
avipad->vids.compression = format;
switch (format) {
case GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'):
avipad->vids.bit_cnt = 16;
break;
case GST_MAKE_FOURCC ('I', '4', '2', '0'):
avipad->vids.bit_cnt = 12;
break;
}
} else {
avipad->vids.bit_cnt = 24;
avipad->vids.compression = 0;
/* find format */
if (!strcmp (mimetype, "video/x-huffyuv")) {
avipad->vids.compression = GST_MAKE_FOURCC ('H', 'F', 'Y', 'U');
} else if (!strcmp (mimetype, "image/jpeg")) {
avipad->vids.compression = GST_MAKE_FOURCC ('M', 'J', 'P', 'G');
} else if (!strcmp (mimetype, "video/x-divx")) {
gint divxversion;
gst_structure_get_int (structure, "divxversion", &divxversion);
switch (divxversion) {
case 3:
avipad->vids.compression = GST_MAKE_FOURCC ('D', 'I', 'V', '3');
break;
case 4:
avipad->vids.compression = GST_MAKE_FOURCC ('D', 'I', 'V', 'X');
break;
case 5:
avipad->vids.compression = GST_MAKE_FOURCC ('D', 'X', '5', '0');
break;
}
} else if (!strcmp (mimetype, "video/x-xvid")) {
avipad->vids.compression = GST_MAKE_FOURCC ('X', 'V', 'I', 'D');
} else if (!strcmp (mimetype, "video/x-3ivx")) {
avipad->vids.compression = GST_MAKE_FOURCC ('3', 'I', 'V', '2');
} else if (gst_structure_has_name (structure, "video/x-msmpeg")) {
gint msmpegversion;
gst_structure_get_int (structure, "msmpegversion", &msmpegversion);
switch (msmpegversion) {
case 41:
avipad->vids.compression = GST_MAKE_FOURCC ('M', 'P', 'G', '4');
break;
case 42:
avipad->vids.compression = GST_MAKE_FOURCC ('M', 'P', '4', '2');
break;
case 43:
avipad->vids.compression = GST_MAKE_FOURCC ('M', 'P', '4', '3');
break;
default:
GST_INFO ("unhandled msmpegversion : %d, fall back to fourcc=MPEG",
msmpegversion);
avipad->vids.compression = GST_MAKE_FOURCC ('M', 'P', 'E', 'G');
break;
}
} else if (!strcmp (mimetype, "video/x-dv")) {
avipad->vids.compression = GST_MAKE_FOURCC ('D', 'V', 'S', 'D');
} else if (!strcmp (mimetype, "video/x-h263")) {
avipad->vids.compression = GST_MAKE_FOURCC ('H', '2', '6', '3');
} else if (!strcmp (mimetype, "video/x-h264")) {
avipad->vids.compression = GST_MAKE_FOURCC ('H', '2', '6', '4');
} else if (!strcmp (mimetype, "video/mpeg")) {
gint mpegversion;
gst_structure_get_int (structure, "mpegversion", &mpegversion);
switch (mpegversion) {
case 2:
avipad->vids.compression = GST_MAKE_FOURCC ('M', 'P', 'G', '2');
break;
case 4:
/* mplayer/ffmpeg might not work with DIVX, but with FMP4 */
avipad->vids.compression = GST_MAKE_FOURCC ('D', 'I', 'V', 'X');
break;
default:
GST_INFO ("unhandled mpegversion : %d, fall back to fourcc=MPEG",
mpegversion);
avipad->vids.compression = GST_MAKE_FOURCC ('M', 'P', 'E', 'G');
break;
}
} else if (!strcmp (mimetype, "video/x-dirac")) {
avipad->vids.compression = GST_MAKE_FOURCC ('d', 'r', 'a', 'c');
} else if (!strcmp (mimetype, "video/x-wmv")) {
gint wmvversion;
if (gst_structure_get_int (structure, "wmvversion", &wmvversion)) {
switch (wmvversion) {
case 1:
avipad->vids.compression = GST_MAKE_FOURCC ('W', 'M', 'V', '1');
break;
case 2:
avipad->vids.compression = GST_MAKE_FOURCC ('W', 'M', 'V', '2');
break;
case 3:
avipad->vids.compression = GST_MAKE_FOURCC ('W', 'M', 'V', '3');
default:
break;
}
}
} else if (!strcmp (mimetype, "image/x-jpc")) {
avipad->vids.compression = GST_MAKE_FOURCC ('M', 'J', '2', 'C');
}
if (!avipad->vids.compression)
goto refuse_caps;
}
avipad->parent.hdr.fcc_handler = avipad->vids.compression;
avipad->vids.image_size = avipad->vids.height * avipad->vids.width;
/* hm, maybe why avi only handles one stream well ... */
avimux->avi_hdr.width = avipad->vids.width;
avimux->avi_hdr.height = avipad->vids.height;
avimux->avi_hdr.us_frame = 1000000. * avipad->parent.hdr.scale /
avipad->parent.hdr.rate;
gst_object_unref (avimux);
return TRUE;
refuse_caps:
{
GST_WARNING_OBJECT (avimux, "refused caps %" GST_PTR_FORMAT, vscaps);
gst_object_unref (avimux);
return FALSE;
}
}
static GstFlowReturn
gst_avi_mux_audsink_scan_mpeg_audio (GstAviMux * avimux, GstAviPad * avipad,
GstBuffer * buffer)
{
guint8 *data;
guint size;
guint spf;
guint32 header;
gulong layer;
gulong version;
gint lsf, mpg25;
data = GST_BUFFER_DATA (buffer);
size = GST_BUFFER_SIZE (buffer);
if (size < 4)
goto not_parsed;
header = GST_READ_UINT32_BE (data);
if ((header & 0xffe00000) != 0xffe00000)
goto not_parsed;
/* thanks go to mp3parse */
if (header & (1 << 20)) {
lsf = (header & (1 << 19)) ? 0 : 1;
mpg25 = 0;
} else {
lsf = 1;
mpg25 = 1;
}
version = 1 + lsf + mpg25;
layer = 4 - ((header >> 17) & 0x3);
/* see http://www.codeproject.com/audio/MPEGAudioInfo.asp */
if (layer == 1)
spf = 384;
else if (layer == 2)
spf = 1152;
else if (version == 1) {
spf = 1152;
} else {
/* MPEG-2 or "2.5" */
spf = 576;
}
if (G_UNLIKELY (avipad->hdr.scale <= 1))
avipad->hdr.scale = spf;
else if (G_UNLIKELY (avipad->hdr.scale != spf)) {
GST_WARNING_OBJECT (avimux, "input mpeg audio has varying frame size");
goto cbr_fallback;
}
return GST_FLOW_OK;
/* EXITS */
not_parsed:
{
GST_WARNING_OBJECT (avimux, "input mpeg audio is not parsed");
/* fall-through */
}
cbr_fallback:
{
GST_WARNING_OBJECT (avimux, "falling back to CBR muxing");
avipad->hdr.scale = 1;
/* no need to check further */
avipad->hook = NULL;
return GST_FLOW_OK;
}
}
static void
gst_avi_mux_audsink_set_fields (GstAviMux * avimux, GstAviAudioPad * avipad)
{
if (avipad->parent.hdr.scale > 1) {
/* vbr case: fixed duration per frame/chunk */
avipad->parent.hdr.rate = avipad->auds.rate;
avipad->parent.hdr.samplesize = 0;
/* FIXME ?? some rumours say this should be largest audio chunk size */
avipad->auds.blockalign = avipad->parent.hdr.scale;
} else {
/* by spec, hdr.rate is av_bps related, is calculated that way in stop_file,
* and reduces to sample rate in PCM like cases */
avipad->parent.hdr.rate = avipad->auds.av_bps / avipad->auds.blockalign;
avipad->parent.hdr.samplesize = avipad->auds.blockalign;
avipad->parent.hdr.scale = 1;
}
}
static gboolean
gst_avi_mux_audsink_set_caps (GstPad * pad, GstCaps * vscaps)
{
GstAviMux *avimux;
GstAviAudioPad *avipad;
GstAviCollectData *collect_pad;
GstStructure *structure;
const gchar *mimetype;
const GValue *codec_data;
gint channels, rate;
avimux = GST_AVI_MUX (gst_pad_get_parent (pad));
/* find stream data */
collect_pad = (GstAviCollectData *) gst_pad_get_element_private (pad);
g_assert (collect_pad);
avipad = (GstAviAudioPad *) collect_pad->avipad;
g_assert (avipad);
g_assert (!avipad->parent.is_video);
g_assert (avipad->parent.hdr.type == GST_MAKE_FOURCC ('a', 'u', 'd', 's'));
GST_DEBUG_OBJECT (avimux, "%s:%s, caps=%" GST_PTR_FORMAT,
GST_DEBUG_PAD_NAME (pad), vscaps);
structure = gst_caps_get_structure (vscaps, 0);
mimetype = gst_structure_get_name (structure);
/* we want these for all */
if (!gst_structure_get_int (structure, "channels", &channels) ||
!gst_structure_get_int (structure, "rate", &rate)) {
goto refuse_caps;
}
avipad->auds.channels = channels;
avipad->auds.rate = rate;
/* codec initialization data, if any */
codec_data = gst_structure_get_value (structure, "codec_data");
if (codec_data) {
avipad->auds_codec_data = gst_value_get_buffer (codec_data);
gst_buffer_ref (avipad->auds_codec_data);
/* keep global track of size */
avimux->codec_data_size += GST_BUFFER_SIZE (avipad->auds_codec_data);
}
if (!strcmp (mimetype, "audio/x-raw-int")) {
gint width, depth;
gboolean signedness;
avipad->auds.format = GST_RIFF_WAVE_FORMAT_PCM;
if (!gst_structure_get_int (structure, "width", &width) ||
!gst_structure_get_int (structure, "depth", &depth) ||
!gst_structure_get_boolean (structure, "signed", &signedness)) {
GST_DEBUG_OBJECT (avimux,
"broken caps, width/depth/signed field missing");
goto refuse_caps;
}
/* no clear place to put different values for these while keeping to spec */
if (width != depth) {
GST_DEBUG_OBJECT (avimux, "width must be same as depth!");
goto refuse_caps;
}
/* because that's the way the caps will be recreated from riff data */
if ((width == 8 && signedness) || (width == 16 && !signedness)) {
GST_DEBUG_OBJECT (avimux,
"8-bit PCM must be unsigned, 16-bit PCM signed");
goto refuse_caps;
}
avipad->auds.blockalign = width;
avipad->auds.size = (width == 8) ? 8 : depth;
/* set some more info straight */
avipad->auds.blockalign /= 8;
avipad->auds.blockalign *= avipad->auds.channels;
avipad->auds.av_bps = avipad->auds.blockalign * avipad->auds.rate;
} else {
avipad->auds.format = 0;
/* set some defaults */
avipad->auds.blockalign = 1;
avipad->auds.av_bps = 0;
avipad->auds.size = 16;
if (!strcmp (mimetype, "audio/mpeg")) {
gint mpegversion;
gst_structure_get_int (structure, "mpegversion", &mpegversion);
switch (mpegversion) {
case 1:{
gint layer = 3;
gboolean parsed = FALSE;
gst_structure_get_int (structure, "layer", &layer);
gst_structure_get_boolean (structure, "parsed", &parsed);
switch (layer) {
case 3:
avipad->auds.format = GST_RIFF_WAVE_FORMAT_MPEGL3;
break;
case 1:
case 2:
avipad->auds.format = GST_RIFF_WAVE_FORMAT_MPEGL12;
break;
}
if (parsed) {
/* treat as VBR, should also cover CBR case;
* setup hook to parse frame header and determine spf */
avipad->parent.hook = gst_avi_mux_audsink_scan_mpeg_audio;
} else {
GST_WARNING_OBJECT (avimux, "unparsed MPEG audio input (?), "
"doing CBR muxing");
}
break;
}
case 4:
{
GstBuffer *codec_data_buf = avipad->auds_codec_data;
guint codec;
/* vbr case needs some special handling */
if (!codec_data_buf || GST_BUFFER_SIZE (codec_data_buf) < 2) {
GST_WARNING_OBJECT (avimux, "no (valid) codec_data for AAC audio");
break;
}
avipad->auds.format = GST_RIFF_WAVE_FORMAT_AAC;
/* need to determine frame length */
codec = GST_READ_UINT16_BE (GST_BUFFER_DATA (codec_data_buf));
avipad->parent.hdr.scale = (codec & 0x4) ? 960 : 1024;
break;
}
}
} else if (!strcmp (mimetype, "audio/x-vorbis")) {
avipad->auds.format = GST_RIFF_WAVE_FORMAT_VORBIS3;
} else if (!strcmp (mimetype, "audio/x-ac3")) {
avipad->auds.format = GST_RIFF_WAVE_FORMAT_A52;
} else if (!strcmp (mimetype, "audio/x-alaw")) {
avipad->auds.format = GST_RIFF_WAVE_FORMAT_ALAW;
avipad->auds.size = 8;
avipad->auds.blockalign = avipad->auds.channels;
avipad->auds.av_bps = avipad->auds.blockalign * avipad->auds.rate;
} else if (!strcmp (mimetype, "audio/x-mulaw")) {
avipad->auds.format = GST_RIFF_WAVE_FORMAT_MULAW;
avipad->auds.size = 8;
avipad->auds.blockalign = avipad->auds.channels;
avipad->auds.av_bps = avipad->auds.blockalign * avipad->auds.rate;
} else if (!strcmp (mimetype, "audio/x-wma")) {
gint version;
gint bitrate;
gint block_align;
if (gst_structure_get_int (structure, "wmaversion", &version)) {
switch (version) {
case 1:
avipad->auds.format = GST_RIFF_WAVE_FORMAT_WMAV1;
break;
case 2:
avipad->auds.format = GST_RIFF_WAVE_FORMAT_WMAV2;
break;
default:
break;
}
}
if (avipad->auds.format != 0) {
if (gst_structure_get_int (structure, "block_align", &block_align)) {
avipad->auds.blockalign = block_align;
}
if (gst_structure_get_int (structure, "bitrate", &bitrate)) {
avipad->auds.av_bps = bitrate / 8;
}
}
}
}
if (!avipad->auds.format)
goto refuse_caps;
gst_avi_mux_audsink_set_fields (avimux, avipad);
gst_object_unref (avimux);
return TRUE;
refuse_caps:
{
GST_WARNING_OBJECT (avimux, "refused caps %" GST_PTR_FORMAT, vscaps);
gst_object_unref (avimux);
return FALSE;
}
}
static GstPad *
gst_avi_mux_request_new_pad (GstElement * element,
GstPadTemplate * templ, const gchar * req_name)
{
GstAviMux *avimux;
GstPad *newpad;
GstAviPad *avipad;
GstElementClass *klass;
g_return_val_if_fail (templ != NULL, NULL);
if (templ->direction != GST_PAD_SINK)
goto wrong_direction;
g_return_val_if_fail (GST_IS_AVI_MUX (element), NULL);
avimux = GST_AVI_MUX (element);
if (!avimux->write_header)
goto too_late;
klass = GST_ELEMENT_GET_CLASS (element);
if (templ == gst_element_class_get_pad_template (klass, "audio_%d")) {
gchar *name;
/* setup pad */
name = g_strdup_printf ("audio_%02d", avimux->audio_pads);
GST_DEBUG_OBJECT (avimux, "adding new pad: %s", name);
newpad = gst_pad_new_from_template (templ, name);
g_free (name);
gst_pad_set_setcaps_function (newpad,
GST_DEBUG_FUNCPTR (gst_avi_mux_audsink_set_caps));
/* init pad specific data */
avipad = g_malloc0 (sizeof (GstAviAudioPad));
avipad->is_video = FALSE;
avipad->hdr.type = GST_MAKE_FOURCC ('a', 'u', 'd', 's');
avimux->audio_pads++;
/* audio goes last */
avimux->sinkpads = g_slist_append (avimux->sinkpads, avipad);
} else if (templ == gst_element_class_get_pad_template (klass, "video_%d")) {
/* though streams are pretty generic and relatively self-contained,
* some video info goes in a single avi header -and therefore mux struct-
* so video restricted to one stream */
if (avimux->video_pads > 0)
return NULL;
/* setup pad */
GST_DEBUG_OBJECT (avimux, "adding new pad: video_00");
newpad = gst_pad_new_from_template (templ, "video_00");
gst_pad_set_setcaps_function (newpad,
GST_DEBUG_FUNCPTR (gst_avi_mux_vidsink_set_caps));
avipad = g_malloc0 (sizeof (GstAviVideoPad));
/* init pad specific data */
avipad->is_video = TRUE;
avipad->hdr.type = GST_MAKE_FOURCC ('v', 'i', 'd', 's');
avimux->video_pads++;
/* video goes first */
avimux->sinkpads = g_slist_prepend (avimux->sinkpads, avipad);
} else
goto wrong_template;
avipad->collect = gst_collect_pads_add_pad (avimux->collect,
newpad, sizeof (GstAviCollectData));
((GstAviCollectData *) (avipad->collect))->avipad = avipad;
/* FIXME: hacked way to override/extend the event function of
* GstCollectPads; because it sets its own event function giving the
* element no access to events */
avimux->collect_event = (GstPadEventFunction) GST_PAD_EVENTFUNC (newpad);
gst_pad_set_event_function (newpad,
GST_DEBUG_FUNCPTR (gst_avi_mux_handle_event));
gst_element_add_pad (element, newpad);
return newpad;
/* ERRORS */
wrong_direction:
{
g_warning ("avimux: request pad that is not a SINK pad\n");
return NULL;
}
too_late:
{
g_warning ("avimux: request pad cannot be added after streaming started\n");
return NULL;
}
wrong_template:
{
g_warning ("avimuxx: this is not our template!\n");
return NULL;
}
}
static void
gst_avi_mux_release_pad (GstElement * element, GstPad * pad)
{
GstAviMux *avimux = GST_AVI_MUX (element);
GSList *node;
node = avimux->sinkpads;
while (node) {
GstAviPad *avipad = (GstAviPad *) node->data;
if (avipad->collect->pad == pad) {
/* pad count should not be adjusted,
* as it also represent number of streams present */
avipad->collect = NULL;
GST_DEBUG_OBJECT (avimux, "removed pad '%s'", GST_PAD_NAME (pad));
gst_collect_pads_remove_pad (avimux->collect, pad);
gst_element_remove_pad (element, pad);
/* if not started yet, we can remove any sign this pad ever existed */
/* in this case _start will take care of the real pad count */
if (avimux->write_header) {
avimux->sinkpads = g_slist_remove (avimux->sinkpads, avipad);
gst_avi_mux_pad_reset (avipad, TRUE);
g_free (avipad);
}
return;
}
node = node->next;
}
g_warning ("Unknown pad %s", GST_PAD_NAME (pad));
}
/* maybe some of these functions should be moved to riff.h? */
/* DISCLAIMER: this function is fairly ugly. So be it (i.e. it makes the rest easier)
* so is this struct */
typedef struct _GstMarkedBuffer
{
guint *highmark;
GstBuffer *buffer;
} GstMarkedBuffer;
static void
gst_avi_mux_write_tag (const GstTagList * list, const gchar * tag,
gpointer data)
{
const struct
{
guint32 fcc;
gchar *tag;
} rifftags[] = {
{
GST_RIFF_INFO_IARL, GST_TAG_LOCATION}, {
GST_RIFF_INFO_IART, GST_TAG_ARTIST}, {
GST_RIFF_INFO_ICMT, GST_TAG_COMMENT}, {
GST_RIFF_INFO_ICOP, GST_TAG_COPYRIGHT}, {
GST_RIFF_INFO_ICRD, GST_TAG_DATE}, {
GST_RIFF_INFO_IGNR, GST_TAG_GENRE}, {
GST_RIFF_INFO_IKEY, GST_TAG_KEYWORDS}, {
GST_RIFF_INFO_INAM, GST_TAG_TITLE}, {
GST_RIFF_INFO_ISFT, GST_TAG_ENCODER}, {
GST_RIFF_INFO_ISRC, GST_TAG_ISRC}, {
0, NULL}
};
gint n, len, plen;
GstBuffer *buf = ((GstMarkedBuffer *) data)->buffer;
guint *highmark = ((GstMarkedBuffer *) data)->highmark;
guint8 *buffdata = GST_BUFFER_DATA (buf) + *highmark;
gchar *str;
for (n = 0; rifftags[n].fcc != 0; n++) {
if (!strcmp (rifftags[n].tag, tag) &&
gst_tag_list_get_string (list, tag, &str) && str) {
len = strlen (str);
plen = len + 1;
if (plen & 1)
plen++;
if (GST_BUFFER_SIZE (buf) >= *highmark + 8 + plen) {
GST_WRITE_UINT32_LE (buffdata, rifftags[n].fcc);
GST_WRITE_UINT32_LE (buffdata + 4, len + 1);
memcpy (buffdata + 8, str, len);
buffdata[8 + len] = 0;
*highmark += 8 + plen;
GST_DEBUG ("writing tag in buffer %p, highmark at %d", buf, *highmark);
}
g_free (str);
break;
}
}
}
#define ODML_SUPERINDEX_SIZE \
(32 + GST_AVI_SUPERINDEX_COUNT * sizeof (gst_avi_superindex_entry))
static GstBuffer *
gst_avi_mux_riff_get_avi_header (GstAviMux * avimux)
{
const GstTagList *tags;
GstBuffer *buffer;
guint8 *buffdata;
guint size = 0;
guint highmark = 0;
/* pointer to list size field */
guint8 *riff_size, *hdrl_size;
GSList *node;
GST_DEBUG_OBJECT (avimux, "creating avi header, data_size %u, idx_size %u",
avimux->data_size, avimux->idx_size);
if (avimux->tags_snap)
tags = avimux->tags_snap;
else {
/* need to make snapshot of current state of tags to ensure the same set
* is used next time around during header rewrite at the end */
tags = gst_tag_setter_get_tag_list (GST_TAG_SETTER (avimux));
if (tags)
tags = avimux->tags_snap = gst_tag_list_copy (tags);
}
if (tags) {
/* that should be the strlen of all tags + header sizes
* not all of tags end up in a avi, still this is a good estimate
*/
gchar *str = gst_structure_to_string (tags);
size += strlen (str) + 8 * gst_structure_n_fields (tags);
g_free (str);
}
/* allocate the buffer, starting with some wild/safe upper bound */
size += avimux->codec_data_size + 100 + sizeof (gst_riff_avih)
+ (g_slist_length (avimux->sinkpads) * (100 + sizeof (gst_riff_strh_full)
+ sizeof (gst_riff_strf_vids)
+ sizeof (gst_riff_vprp)
+ sizeof (gst_riff_vprp_video_field_desc) * 2
+ sizeof (gst_riff_strf_auds) + 2 + ODML_SUPERINDEX_SIZE));
buffer = gst_buffer_new_and_alloc (size);
buffdata = GST_BUFFER_DATA (buffer);
highmark = 0;
GST_DEBUG_OBJECT (avimux, "creating buffer %p, size %d, highmark at 0",
buffer, GST_BUFFER_SIZE (buffer));
/* avi header metadata */
memcpy (buffdata + 0, "RIFF", 4);
/* fill in RIFF size later */
riff_size = buffdata + 4;
memcpy (buffdata + 8, "AVI ", 4);
memcpy (buffdata + 12, "LIST", 4);
/* fill in header size later */
hdrl_size = buffdata + 16;
memcpy (buffdata + 20, "hdrl", 4);
memcpy (buffdata + 24, "avih", 4);
GST_WRITE_UINT32_LE (buffdata + 28, sizeof (gst_riff_avih));
buffdata += 32;
highmark += 32;
/* the AVI header itself */
GST_WRITE_UINT32_LE (buffdata + 0, avimux->avi_hdr.us_frame);
GST_WRITE_UINT32_LE (buffdata + 4, avimux->avi_hdr.max_bps);
GST_WRITE_UINT32_LE (buffdata + 8, avimux->avi_hdr.pad_gran);
GST_WRITE_UINT32_LE (buffdata + 12, avimux->avi_hdr.flags);
GST_WRITE_UINT32_LE (buffdata + 16, avimux->avi_hdr.tot_frames);
GST_WRITE_UINT32_LE (buffdata + 20, avimux->avi_hdr.init_frames);
GST_WRITE_UINT32_LE (buffdata + 24, avimux->avi_hdr.streams);
GST_WRITE_UINT32_LE (buffdata + 28, avimux->avi_hdr.bufsize);
GST_WRITE_UINT32_LE (buffdata + 32, avimux->avi_hdr.width);
GST_WRITE_UINT32_LE (buffdata + 36, avimux->avi_hdr.height);
GST_WRITE_UINT32_LE (buffdata + 40, avimux->avi_hdr.scale);
GST_WRITE_UINT32_LE (buffdata + 44, avimux->avi_hdr.rate);
GST_WRITE_UINT32_LE (buffdata + 48, avimux->avi_hdr.start);
GST_WRITE_UINT32_LE (buffdata + 52, avimux->avi_hdr.length);
buffdata += 56;
highmark += 56;
/* stream data */
node = avimux->sinkpads;
while (node) {
GstAviPad *avipad = (GstAviPad *) node->data;
GstAviVideoPad *vidpad = (GstAviVideoPad *) avipad;
GstAviAudioPad *audpad = (GstAviAudioPad *) avipad;
guint codec_size = 0, strl_size = 0, vprp_size = 0;
if (avipad->is_video) {
if (vidpad->vids_codec_data)
codec_size = GST_BUFFER_SIZE (vidpad->vids_codec_data);
strl_size = sizeof (gst_riff_strh_full) + sizeof (gst_riff_strf_vids)
+ GST_ROUND_UP_2 (codec_size) + 4 * 5 + ODML_SUPERINDEX_SIZE;
if (vidpad->vprp.aspect) {
/* let's be on the safe side */
vidpad->vprp.fields = MIN (vidpad->vprp.fields,
GST_RIFF_VPRP_VIDEO_FIELDS);
vprp_size = G_STRUCT_OFFSET (gst_riff_vprp, field_info)
+ (vidpad->vprp.fields * sizeof (gst_riff_vprp_video_field_desc));
strl_size += 4 * 2 + vprp_size;
}
} else {
if (audpad->auds_codec_data)
codec_size = GST_BUFFER_SIZE (audpad->auds_codec_data);
/* +2 is codec_size field, not part of gst_riff_strf_auds */
strl_size = sizeof (gst_riff_strh_full) + sizeof (gst_riff_strf_auds) + 2
+ GST_ROUND_UP_2 (codec_size) + 4 * 5 + ODML_SUPERINDEX_SIZE;
}
/* stream list metadata */
memcpy (buffdata + 0, "LIST", 4);
GST_WRITE_UINT32_LE (buffdata + 4, strl_size);
memcpy (buffdata + 8, "strl", 4);
/* generic header */
memcpy (buffdata + 12, "strh", 4);
GST_WRITE_UINT32_LE (buffdata + 16, sizeof (gst_riff_strh_full));
/* the actual header */
GST_WRITE_UINT32_LE (buffdata + 20, avipad->hdr.type);
GST_WRITE_UINT32_LE (buffdata + 24, avipad->hdr.fcc_handler);
GST_WRITE_UINT32_LE (buffdata + 28, avipad->hdr.flags);
GST_WRITE_UINT32_LE (buffdata + 32, avipad->hdr.priority);
GST_WRITE_UINT32_LE (buffdata + 36, avipad->hdr.init_frames);
GST_WRITE_UINT32_LE (buffdata + 40, avipad->hdr.scale);
GST_WRITE_UINT32_LE (buffdata + 44, avipad->hdr.rate);
GST_WRITE_UINT32_LE (buffdata + 48, avipad->hdr.start);
GST_WRITE_UINT32_LE (buffdata + 52, avipad->hdr.length);
GST_WRITE_UINT32_LE (buffdata + 56, avipad->hdr.bufsize);
GST_WRITE_UINT32_LE (buffdata + 60, avipad->hdr.quality);
GST_WRITE_UINT32_LE (buffdata + 64, avipad->hdr.samplesize);
GST_WRITE_UINT16_LE (buffdata + 68, 0);
GST_WRITE_UINT16_LE (buffdata + 70, 0);
GST_WRITE_UINT16_LE (buffdata + 72, 0);
GST_WRITE_UINT16_LE (buffdata + 74, 0);
buffdata += 76;
highmark += 76;
if (avipad->is_video) {
/* the video header */
memcpy (buffdata + 0, "strf", 4);
GST_WRITE_UINT32_LE (buffdata + 4,
sizeof (gst_riff_strf_vids) + codec_size);
/* the actual header */
GST_WRITE_UINT32_LE (buffdata + 8, vidpad->vids.size + codec_size);
GST_WRITE_UINT32_LE (buffdata + 12, vidpad->vids.width);
GST_WRITE_UINT32_LE (buffdata + 16, vidpad->vids.height);
GST_WRITE_UINT16_LE (buffdata + 20, vidpad->vids.planes);
GST_WRITE_UINT16_LE (buffdata + 22, vidpad->vids.bit_cnt);
GST_WRITE_UINT32_LE (buffdata + 24, vidpad->vids.compression);
GST_WRITE_UINT32_LE (buffdata + 28, vidpad->vids.image_size);
GST_WRITE_UINT32_LE (buffdata + 32, vidpad->vids.xpels_meter);
GST_WRITE_UINT32_LE (buffdata + 36, vidpad->vids.ypels_meter);
GST_WRITE_UINT32_LE (buffdata + 40, vidpad->vids.num_colors);
GST_WRITE_UINT32_LE (buffdata + 44, vidpad->vids.imp_colors);
buffdata += 48;
highmark += 48;
/* include codec data, if any */
if (codec_size) {
memcpy (buffdata, GST_BUFFER_DATA (vidpad->vids_codec_data),
codec_size);
buffdata += codec_size;
highmark += codec_size;
}
/* padding */
if (highmark & 0x1) {
highmark++;
buffdata++;
}
/* add video property data, mainly for aspect ratio, if any */
if (vprp_size) {
gint f;
/* the vprp header */
memcpy (buffdata + 0, "vprp", 4);
GST_WRITE_UINT32_LE (buffdata + 4, vprp_size);
/* the actual data */
GST_WRITE_UINT32_LE (buffdata + 8, vidpad->vprp.format_token);
GST_WRITE_UINT32_LE (buffdata + 12, vidpad->vprp.standard);
GST_WRITE_UINT32_LE (buffdata + 16, vidpad->vprp.vert_rate);
GST_WRITE_UINT32_LE (buffdata + 20, vidpad->vprp.hor_t_total);
GST_WRITE_UINT32_LE (buffdata + 24, vidpad->vprp.vert_lines);
GST_WRITE_UINT32_LE (buffdata + 28, vidpad->vprp.aspect);
GST_WRITE_UINT32_LE (buffdata + 32, vidpad->vprp.width);
GST_WRITE_UINT32_LE (buffdata + 36, vidpad->vprp.height);
GST_WRITE_UINT32_LE (buffdata + 40, vidpad->vprp.fields);
buffdata += 44;
highmark += 44;
for (f = 0; f < vidpad->vprp.fields; ++f) {
gst_riff_vprp_video_field_desc *fd;
fd = &(vidpad->vprp.field_info[f]);
GST_WRITE_UINT32_LE (buffdata + 0, fd->compressed_bm_height);
GST_WRITE_UINT32_LE (buffdata + 4, fd->compressed_bm_width);
GST_WRITE_UINT32_LE (buffdata + 8, fd->valid_bm_height);
GST_WRITE_UINT32_LE (buffdata + 12, fd->valid_bm_width);
GST_WRITE_UINT32_LE (buffdata + 16, fd->valid_bm_x_offset);
GST_WRITE_UINT32_LE (buffdata + 20, fd->valid_bm_y_offset);
GST_WRITE_UINT32_LE (buffdata + 24, fd->video_x_t_offset);
GST_WRITE_UINT32_LE (buffdata + 28, fd->video_y_start);
buffdata += 32;
highmark += 32;
}
}
} else {
/* the audio header */
memcpy (buffdata + 0, "strf", 4);
GST_WRITE_UINT32_LE (buffdata + 4,
sizeof (gst_riff_strf_auds) + 2 + codec_size);
/* the actual header */
GST_WRITE_UINT16_LE (buffdata + 8, audpad->auds.format);
GST_WRITE_UINT16_LE (buffdata + 10, audpad->auds.channels);
GST_WRITE_UINT32_LE (buffdata + 12, audpad->auds.rate);
GST_WRITE_UINT32_LE (buffdata + 16, audpad->auds.av_bps);
GST_WRITE_UINT16_LE (buffdata + 20, audpad->auds.blockalign);
GST_WRITE_UINT16_LE (buffdata + 22, audpad->auds.size);
GST_WRITE_UINT16_LE (buffdata + 24, codec_size);
buffdata += 26;
highmark += 26;
/* include codec data, if any */
if (codec_size) {
memcpy (buffdata, GST_BUFFER_DATA (audpad->auds_codec_data),
codec_size);
buffdata += codec_size;
highmark += codec_size;
}
/* padding */
if (highmark & 0x1) {
highmark++;
buffdata++;
}
}
/* odml superindex chunk */
if (avipad->idx_index > 0)
memcpy (buffdata, "indx", 4);
else
memcpy (buffdata, "JUNK", 4);
GST_WRITE_UINT32_LE (buffdata + 4, ODML_SUPERINDEX_SIZE - 8); /* chunk size */
GST_WRITE_UINT16_LE (buffdata + 8, 4); /* bytes per entry */
buffdata[10] = 0; /* index subtype */
buffdata[11] = GST_AVI_INDEX_OF_INDEXES; /* index type */
GST_WRITE_UINT32_LE (buffdata + 12, avipad->idx_index); /* entries in use */
memcpy (buffdata + 16, avipad->tag, 4); /* stream id */
GST_WRITE_UINT32_LE (buffdata + 20, 0); /* reserved */
GST_WRITE_UINT32_LE (buffdata + 24, 0); /* reserved */
GST_WRITE_UINT32_LE (buffdata + 28, 0); /* reserved */
memcpy (buffdata + 32, avipad->idx,
GST_AVI_SUPERINDEX_COUNT * sizeof (gst_avi_superindex_entry));
buffdata += ODML_SUPERINDEX_SIZE;
highmark += ODML_SUPERINDEX_SIZE;
node = node->next;
}
if (avimux->video_pads > 0) {
/* odml header */
memcpy (buffdata + 0, "LIST", 4);
GST_WRITE_UINT32_LE (buffdata + 4, sizeof (guint32) + 4 * 3);
memcpy (buffdata + 8, "odml", 4);
memcpy (buffdata + 12, "dmlh", 4);
GST_WRITE_UINT32_LE (buffdata + 16, sizeof (guint32));
GST_WRITE_UINT32_LE (buffdata + 20, avimux->total_frames);
buffdata += 24;
highmark += 24;
}
GST_WRITE_UINT32_LE (hdrl_size, (guint32) (buffdata - hdrl_size) - 4);
/* tags */
if (tags) {
guint8 *ptr;
guint startsize;
GstMarkedBuffer data = { &highmark, buffer };
memcpy (buffdata + 0, "LIST", 4);
ptr = buffdata + 4; /* fill in later */
startsize = highmark + 4;
memcpy (buffdata + 8, "INFO", 4);
buffdata += 12;
highmark += 12;
/* 12 bytes is needed for data header */
GST_BUFFER_SIZE (buffer) -= 12;
gst_tag_list_foreach (tags, gst_avi_mux_write_tag, &data);
GST_BUFFER_SIZE (buffer) += 12;
buffdata = GST_BUFFER_DATA (buffer) + highmark;
if (highmark - startsize - 4 == 4) {
/* no tags writen, remove the empty INFO LIST as it is useless
* and prevents playback in vlc */
highmark -= 12;
buffdata = GST_BUFFER_DATA (buffer) + highmark;
/* no need to erase the writen data, it will be overwriten anyway */
} else {
/* update list size */
GST_WRITE_UINT32_LE (ptr, highmark - startsize - 4);
}
}
/* avi data header */
memcpy (buffdata + 0, "LIST", 4);
GST_WRITE_UINT32_LE (buffdata + 4, avimux->data_size);
memcpy (buffdata + 8, "movi", 4);
buffdata += 12;
highmark += 12;
/* finally we can fill in the RIFF size */
/* note that riff only counts the first avi chunk */
GST_WRITE_UINT32_LE (riff_size, (guint32) (buffdata - riff_size - 4) /* header and movi tags */
+avimux->idx_size + avimux->data_size - 4); /* movi data and index */
{ /* only the part that is filled in actually makes up the header
* unref the parent as we only need this part from now on */
GstBuffer *subbuffer = gst_buffer_create_sub (buffer, 0, highmark);
gst_buffer_unref (buffer);
return subbuffer;
}
}
static GstBuffer *
gst_avi_mux_riff_get_avix_header (guint32 datax_size)
{
GstBuffer *buffer;
guint8 *buffdata;
buffer = gst_buffer_new_and_alloc (24);
buffdata = GST_BUFFER_DATA (buffer);
memcpy (buffdata + 0, "RIFF", 4);
GST_WRITE_UINT32_LE (buffdata + 4, datax_size + 3 * 4);
memcpy (buffdata + 8, "AVIX", 4);
memcpy (buffdata + 12, "LIST", 4);
GST_WRITE_UINT32_LE (buffdata + 16, datax_size);
memcpy (buffdata + 20, "movi", 4);
return buffer;
}
static inline GstBuffer *
gst_avi_mux_riff_get_header (GstAviPad * avipad, guint32 video_frame_size)
{
GstBuffer *buffer;
guint8 *buffdata;
buffer = gst_buffer_new_and_alloc (8);
buffdata = GST_BUFFER_DATA (buffer);
memcpy (buffdata + 0, avipad->tag, 4);
GST_WRITE_UINT32_LE (buffdata + 4, video_frame_size);
return buffer;
}
/* write an odml index chunk in the movi list */
static GstFlowReturn
gst_avi_mux_write_avix_index (GstAviMux * avimux, GstAviPad * avipad,
gchar * code, gchar * chunk, gst_avi_superindex_entry * super_index,
gint * super_index_count)
{
GstFlowReturn res;
GstBuffer *buffer;
guint8 *buffdata, *data;
gst_riff_index_entry *entry;
gint i;
guint32 size, entry_count;
gboolean is_pcm = FALSE;
guint32 pcm_samples = 0;
/* check if it is pcm */
if (avipad && !avipad->is_video) {
GstAviAudioPad *audiopad = (GstAviAudioPad *) avipad;
if (audiopad->auds.format == GST_RIFF_WAVE_FORMAT_PCM) {
pcm_samples = audiopad->samples;
is_pcm = TRUE;
}
}
/* allocate the maximum possible */
buffer = gst_buffer_new_and_alloc (32 + 8 * avimux->idx_index);
buffdata = GST_BUFFER_DATA (buffer);
/* general index chunk info */
memcpy (buffdata + 0, chunk, 4); /* chunk id */
GST_WRITE_UINT32_LE (buffdata + 4, 0); /* chunk size; fill later */
GST_WRITE_UINT16_LE (buffdata + 8, 2); /* index entry is 2 words */
buffdata[10] = 0; /* index subtype */
buffdata[11] = GST_AVI_INDEX_OF_CHUNKS; /* index type: AVI_INDEX_OF_CHUNKS */
GST_WRITE_UINT32_LE (buffdata + 12, 0); /* entries in use; fill later */
memcpy (buffdata + 16, code, 4); /* stream to which index refers */
GST_WRITE_UINT64_LE (buffdata + 20, avimux->avix_start); /* base offset */
GST_WRITE_UINT32_LE (buffdata + 28, 0); /* reserved */
buffdata += 32;
/* now the actual index entries */
i = avimux->idx_index;
entry = avimux->idx;
while (i > 0) {
if (memcmp (&entry->id, code, 4) == 0) {
/* enter relative offset to the data (!) */
GST_WRITE_UINT32_LE (buffdata, GUINT32_FROM_LE (entry->offset) + 8);
/* msb is set if not (!) keyframe */
GST_WRITE_UINT32_LE (buffdata + 4, GUINT32_FROM_LE (entry->size)
| (GUINT32_FROM_LE (entry->flags)
& GST_RIFF_IF_KEYFRAME ? 0 : 1U << 31));
buffdata += 8;
}
i--;
entry++;
}
/* ok, now we know the size and no of entries, fill in where needed */
data = GST_BUFFER_DATA (buffer);
GST_BUFFER_SIZE (buffer) = size = buffdata - data;
GST_WRITE_UINT32_LE (data + 4, size - 8);
entry_count = (size - 32) / 8;
GST_WRITE_UINT32_LE (data + 12, entry_count);
/* decorate and send */
gst_buffer_set_caps (buffer, GST_PAD_CAPS (avimux->srcpad));
if ((res = gst_pad_push (avimux->srcpad, buffer)) != GST_FLOW_OK)
return res;
/* keep track of this in superindex (if room) ... */
if (*super_index_count < GST_AVI_SUPERINDEX_COUNT) {
i = *super_index_count;
super_index[i].offset = GUINT64_TO_LE (avimux->total_data);
super_index[i].size = GUINT32_TO_LE (size);
if (is_pcm) {
super_index[i].duration = GUINT32_TO_LE (pcm_samples);
} else {
super_index[i].duration = GUINT32_TO_LE (entry_count);
}
(*super_index_count)++;
} else
GST_WARNING_OBJECT (avimux, "No more room in superindex of stream %s",
code);
/* ... and in size */
avimux->total_data += size;
if (avimux->is_bigfile)
avimux->datax_size += size;
else
avimux->data_size += size;
return GST_FLOW_OK;
}
/* some other usable functions (thankyou xawtv ;-) ) */
static void
gst_avi_mux_add_index (GstAviMux * avimux, GstAviPad * avipad, guint32 flags,
guint32 size)
{
gchar *code = avipad->tag;
if (avimux->idx_index == avimux->idx_count) {
avimux->idx_count += 256;
avimux->idx =
g_realloc (avimux->idx,
avimux->idx_count * sizeof (gst_riff_index_entry));
}
/* in case of pcm audio, we need to count the number of samples for
* putting in the indx entries */
if (!avipad->is_video) {
GstAviAudioPad *audiopad = (GstAviAudioPad *) avipad;
if (audiopad->auds.format == GST_RIFF_WAVE_FORMAT_PCM) {
audiopad->samples += size / audiopad->auds.blockalign;
}
}
memcpy (&(avimux->idx[avimux->idx_index].id), code, 4);
avimux->idx[avimux->idx_index].flags = GUINT32_TO_LE (flags);
avimux->idx[avimux->idx_index].offset = GUINT32_TO_LE (avimux->idx_offset);
avimux->idx[avimux->idx_index].size = GUINT32_TO_LE (size);
avimux->idx_index++;
}
static GstFlowReturn
gst_avi_mux_write_index (GstAviMux * avimux)
{
GstFlowReturn res;
GstBuffer *buffer;
guint8 *buffdata;
buffer = gst_buffer_new_and_alloc (8);
buffdata = GST_BUFFER_DATA (buffer);
memcpy (buffdata + 0, "idx1", 4);
GST_WRITE_UINT32_LE (buffdata + 4,
avimux->idx_index * sizeof (gst_riff_index_entry));
gst_buffer_set_caps (buffer, GST_PAD_CAPS (avimux->srcpad));
res = gst_pad_push (avimux->srcpad, buffer);
if (res != GST_FLOW_OK)
return res;
buffer = gst_buffer_new ();
GST_BUFFER_SIZE (buffer) = avimux->idx_index * sizeof (gst_riff_index_entry);
GST_BUFFER_DATA (buffer) = (guint8 *) avimux->idx;
GST_BUFFER_MALLOCDATA (buffer) = GST_BUFFER_DATA (buffer);
avimux->idx = NULL; /* will be free()'ed by gst_buffer_unref() */
avimux->total_data += GST_BUFFER_SIZE (buffer) + 8;
gst_buffer_set_caps (buffer, GST_PAD_CAPS (avimux->srcpad));
res = gst_pad_push (avimux->srcpad, buffer);
if (res != GST_FLOW_OK)
return res;
avimux->idx_size += avimux->idx_index * sizeof (gst_riff_index_entry) + 8;
/* update header */
avimux->avi_hdr.flags |= GST_RIFF_AVIH_HASINDEX;
return GST_FLOW_OK;
}
static GstFlowReturn
gst_avi_mux_bigfile (GstAviMux * avimux, gboolean last)
{
GstFlowReturn res = GST_FLOW_OK;
GstBuffer *header;
GstEvent *event;
GSList *node;
/* first some odml standard index chunks in the movi list */
node = avimux->sinkpads;
while (node) {
GstAviPad *avipad = (GstAviPad *) node->data;
node = node->next;
res = gst_avi_mux_write_avix_index (avimux, avipad, avipad->tag,
avipad->idx_tag, avipad->idx, &avipad->idx_index);
if (res != GST_FLOW_OK)
return res;
}
if (avimux->is_bigfile) {
/* search back */
event = gst_event_new_new_segment (FALSE, 1.0, GST_FORMAT_BYTES,
avimux->avix_start, GST_CLOCK_TIME_NONE, avimux->avix_start);
/* if the event succeeds */
gst_pad_push_event (avimux->srcpad, event);
/* rewrite AVIX header */
header = gst_avi_mux_riff_get_avix_header (avimux->datax_size);
gst_buffer_set_caps (header, GST_PAD_CAPS (avimux->srcpad));
res = gst_pad_push (avimux->srcpad, header);
/* go back to current location, at least try */
event = gst_event_new_new_segment (FALSE, 1.0, GST_FORMAT_BYTES,
avimux->total_data, GST_CLOCK_TIME_NONE, avimux->total_data);
gst_pad_push_event (avimux->srcpad, event);
if (res != GST_FLOW_OK)
return res;
} else { /* write a standard index in the first riff chunk */
res = gst_avi_mux_write_index (avimux);
/* the index data/buffer is freed by pushing it */
avimux->idx_count = 0;
if (res != GST_FLOW_OK)
return res;
}
avimux->avix_start = avimux->total_data;
if (last)
return res;
avimux->is_bigfile = TRUE;
avimux->numx_frames = 0;
avimux->datax_size = 4; /* movi tag */
avimux->idx_index = 0;
node = avimux->sinkpads;
while (node) {
GstAviPad *avipad = (GstAviPad *) node->data;
node = node->next;
if (!avipad->is_video) {
GstAviAudioPad *audiopad = (GstAviAudioPad *) avipad;
audiopad->samples = 0;
}
}
header = gst_avi_mux_riff_get_avix_header (0);
avimux->total_data += GST_BUFFER_SIZE (header);
/* avix_start is used as base offset for the odml index chunk */
avimux->idx_offset = avimux->total_data - avimux->avix_start;
gst_buffer_set_caps (header, GST_PAD_CAPS (avimux->srcpad));
return gst_pad_push (avimux->srcpad, header);
}
/* enough header blabla now, let's go on to actually writing the headers */
static GstFlowReturn
gst_avi_mux_start_file (GstAviMux * avimux)
{
GstFlowReturn res;
GstBuffer *header;
GSList *node;
avimux->total_data = 0;
avimux->total_frames = 0;
avimux->data_size = 4; /* movi tag */
avimux->datax_size = 0;
avimux->num_frames = 0;
avimux->numx_frames = 0;
avimux->avix_start = 0;
avimux->idx_index = 0;
avimux->idx_offset = 0; /* see 10 lines below */
avimux->idx_size = 0;
avimux->idx_count = 0;
avimux->idx = NULL;
/* state */
avimux->write_header = FALSE;
avimux->restart = FALSE;
/* init streams, see what we've got */
node = avimux->sinkpads;
avimux->audio_pads = avimux->video_pads = 0;
while (node) {
GstAviPad *avipad = (GstAviPad *) node->data;
node = node->next;
if (!avipad->is_video) {
/* audio stream numbers must start at 1 iff there is a video stream 0;
* request_pad inserts video pad at head of list, so this test suffices */
if (avimux->video_pads)
avimux->audio_pads++;
avipad->tag = g_strdup_printf ("%02uwb", avimux->audio_pads);
avipad->idx_tag = g_strdup_printf ("ix%02u", avimux->audio_pads);
if (!avimux->video_pads)
avimux->audio_pads++;
} else {
avipad->tag = g_strdup_printf ("%02udb", avimux->video_pads);
avipad->idx_tag = g_strdup_printf ("ix%02u", avimux->video_pads++);
}
}
/* let downstream know we think in BYTES and expect to do seeking later on */
gst_pad_push_event (avimux->srcpad,
gst_event_new_new_segment (FALSE, 1.0, GST_FORMAT_BYTES, 0, -1, 0));
/* header */
avimux->avi_hdr.streams = g_slist_length (avimux->sinkpads);
avimux->is_bigfile = FALSE;
header = gst_avi_mux_riff_get_avi_header (avimux);
avimux->total_data += GST_BUFFER_SIZE (header);
gst_buffer_set_caps (header, GST_PAD_CAPS (avimux->srcpad));
res = gst_pad_push (avimux->srcpad, header);
avimux->idx_offset = avimux->total_data;
return res;
}
static GstFlowReturn
gst_avi_mux_stop_file (GstAviMux * avimux)
{
GstFlowReturn res = GST_FLOW_OK;
GstEvent *event;
GstBuffer *header;
GSList *node;
/* if bigfile, rewrite header, else write indexes */
/* don't bail out at once if error, still try to re-write header */
if (avimux->video_pads > 0) {
if (avimux->is_bigfile) {
res = gst_avi_mux_bigfile (avimux, TRUE);
} else {
res = gst_avi_mux_write_index (avimux);
}
}
/* we do our best to make it interleaved at least ... */
if (avimux->audio_pads > 0 && avimux->video_pads > 0)
avimux->avi_hdr.flags |= GST_RIFF_AVIH_ISINTERLEAVED;
/* set rate and everything having to do with that */
avimux->avi_hdr.max_bps = 0;
node = avimux->sinkpads;
while (node) {
GstAviPad *avipad = (GstAviPad *) node->data;
node = node->next;
if (!avipad->is_video) {
GstAviAudioPad *audpad = (GstAviAudioPad *) avipad;
/* calculate bps if needed */
if (!audpad->auds.av_bps) {
if (audpad->audio_time) {
audpad->auds.av_bps =
(GST_SECOND * audpad->audio_size) / audpad->audio_time;
/* round bps to nearest multiple of 8;
* which is much more likely to be the (cbr) bitrate in use;
* which in turn results in better timestamp calculation on playback */
audpad->auds.av_bps = GST_ROUND_UP_8 (audpad->auds.av_bps - 4);
} else {
GST_ELEMENT_WARNING (avimux, STREAM, MUX,
(_("No or invalid input audio, AVI stream will be corrupt.")),
(NULL));
audpad->auds.av_bps = 0;
}
}
gst_avi_mux_audsink_set_fields (avimux, audpad);
avimux->avi_hdr.max_bps += audpad->auds.av_bps;
avipad->hdr.length = gst_util_uint64_scale (audpad->audio_time,
avipad->hdr.rate, avipad->hdr.scale * GST_SECOND);
} else {
GstAviVideoPad *vidpad = (GstAviVideoPad *) avipad;
avimux->avi_hdr.max_bps += ((vidpad->vids.bit_cnt + 7) / 8) *
(1000000. / avimux->avi_hdr.us_frame) * vidpad->vids.image_size;
avipad->hdr.length = avimux->total_frames;
}
}
/* statistics/total_frames/... */
avimux->avi_hdr.tot_frames = avimux->num_frames;
/* seek and rewrite the header */
header = gst_avi_mux_riff_get_avi_header (avimux);
event = gst_event_new_new_segment (FALSE, 1.0, GST_FORMAT_BYTES,
0, GST_CLOCK_TIME_NONE, 0);
gst_pad_push_event (avimux->srcpad, event);
gst_buffer_set_caps (header, GST_PAD_CAPS (avimux->srcpad));
/* the first error survives */
if (res == GST_FLOW_OK)
res = gst_pad_push (avimux->srcpad, header);
else
gst_pad_push (avimux->srcpad, header);
event = gst_event_new_new_segment (FALSE, 1.0, GST_FORMAT_BYTES,
avimux->total_data, GST_CLOCK_TIME_NONE, avimux->total_data);
gst_pad_push_event (avimux->srcpad, event);
avimux->write_header = TRUE;
return res;
}
static GstFlowReturn
gst_avi_mux_restart_file (GstAviMux * avimux)
{
GstFlowReturn res;
if ((res = gst_avi_mux_stop_file (avimux)) != GST_FLOW_OK)
return res;
gst_pad_push_event (avimux->srcpad, gst_event_new_eos ());
return gst_avi_mux_start_file (avimux);
}
/* handle events (search) */
static gboolean
gst_avi_mux_handle_event (GstPad * pad, GstEvent * event)
{
GstAviMux *avimux;
gboolean ret;
avimux = GST_AVI_MUX (gst_pad_get_parent (pad));
switch (GST_EVENT_TYPE (event)) {
case GST_EVENT_TAG:{
GstTagList *list;
GstTagSetter *setter = GST_TAG_SETTER (avimux);
const GstTagMergeMode mode = gst_tag_setter_get_tag_merge_mode (setter);
gst_event_parse_tag (event, &list);
gst_tag_setter_merge_tags (setter, list, mode);
break;
}
default:
break;
}
/* now GstCollectPads can take care of the rest, e.g. EOS */
ret = avimux->collect_event (pad, event);
gst_object_unref (avimux);
return ret;
}
/* send extra 'padding' data */
static GstFlowReturn
gst_avi_mux_send_pad_data (GstAviMux * avimux, gulong num_bytes)
{
GstBuffer *buffer;
buffer = gst_buffer_new_and_alloc (num_bytes);
memset (GST_BUFFER_DATA (buffer), 0, num_bytes);
gst_buffer_set_caps (buffer, GST_PAD_CAPS (avimux->srcpad));
return gst_pad_push (avimux->srcpad, buffer);
}
/* do buffer */
static GstFlowReturn
gst_avi_mux_do_buffer (GstAviMux * avimux, GstAviPad * avipad)
{
GstFlowReturn res;
GstBuffer *data, *header;
gulong total_size, pad_bytes = 0;
guint flags;
data = gst_collect_pads_pop (avimux->collect, avipad->collect);
if (avimux->restart) {
if ((res = gst_avi_mux_restart_file (avimux)) != GST_FLOW_OK)
return res;
}
/* need to restart or start a next avix chunk ? */
if ((avimux->is_bigfile ? avimux->datax_size : avimux->data_size) +
GST_BUFFER_SIZE (data) > GST_AVI_MAX_SIZE) {
if (avimux->enable_large_avi) {
if ((res = gst_avi_mux_bigfile (avimux, FALSE)) != GST_FLOW_OK)
return res;
} else {
if ((res = gst_avi_mux_restart_file (avimux)) != GST_FLOW_OK)
return res;
}
}
/* get header and record some stats */
if (GST_BUFFER_SIZE (data) & 1) {
pad_bytes = 2 - (GST_BUFFER_SIZE (data) & 1);
}
header = gst_avi_mux_riff_get_header (avipad, GST_BUFFER_SIZE (data));
total_size = GST_BUFFER_SIZE (header) + GST_BUFFER_SIZE (data) + pad_bytes;
if (avimux->is_bigfile) {
avimux->datax_size += total_size;
} else {
avimux->data_size += total_size;
}
if (G_UNLIKELY (avipad->hook))
avipad->hook (avimux, avipad, data);
/* the suggested buffer size is the max frame size */
if (avipad->hdr.bufsize < GST_BUFFER_SIZE (data))
avipad->hdr.bufsize = GST_BUFFER_SIZE (data);
if (avipad->is_video) {
avimux->total_frames++;
if (avimux->is_bigfile) {
avimux->numx_frames++;
} else {
avimux->num_frames++;
}
flags = 0x02;
if (!GST_BUFFER_FLAG_IS_SET (data, GST_BUFFER_FLAG_DELTA_UNIT))
flags |= 0x10;
} else {
GstAviAudioPad *audpad = (GstAviAudioPad *) avipad;
flags = 0;
audpad->audio_size += GST_BUFFER_SIZE (data);
audpad->audio_time += GST_BUFFER_DURATION (data);
}
gst_avi_mux_add_index (avimux, avipad, flags, GST_BUFFER_SIZE (data));
/* prepare buffers for sending */
gst_buffer_set_caps (header, GST_PAD_CAPS (avimux->srcpad));
data = gst_buffer_make_metadata_writable (data);
gst_buffer_set_caps (data, GST_PAD_CAPS (avimux->srcpad));
GST_LOG_OBJECT (avimux, "pushing buffers: head, data");
if ((res = gst_pad_push (avimux->srcpad, header)) != GST_FLOW_OK)
return res;
if ((res = gst_pad_push (avimux->srcpad, data)) != GST_FLOW_OK)
return res;
if (pad_bytes) {
if ((res = gst_avi_mux_send_pad_data (avimux, pad_bytes)) != GST_FLOW_OK)
return res;
}
/* if any push above fails, we're in trouble with file consistency anyway */
avimux->total_data += total_size;
avimux->idx_offset += total_size;
return res;
}
/* pick the oldest buffer from the pads and push it */
static GstFlowReturn
gst_avi_mux_do_one_buffer (GstAviMux * avimux)
{
GstAviPad *avipad, *best_pad;
GSList *node;
GstBuffer *buffer;
GstClockTime time, best_time, delay;
node = avimux->sinkpads;
best_pad = NULL;
best_time = GST_CLOCK_TIME_NONE;
for (; node; node = node->next) {
avipad = (GstAviPad *) node->data;
if (!avipad->collect)
continue;
buffer = gst_collect_pads_peek (avimux->collect, avipad->collect);
if (!buffer)
continue;
time = GST_BUFFER_TIMESTAMP (buffer);
gst_buffer_unref (buffer);
delay = avipad->is_video ? GST_SECOND / 2 : 0;
/* invalid timestamp buffers pass first,
* these are probably initialization buffers */
if (best_pad == NULL || !GST_CLOCK_TIME_IS_VALID (time)
|| (GST_CLOCK_TIME_IS_VALID (best_time) && time + delay < best_time)) {
best_pad = avipad;
best_time = time + delay;
}
}
if (best_pad) {
GST_LOG_OBJECT (avimux, "selected pad %s with time %" GST_TIME_FORMAT,
GST_PAD_NAME (best_pad->collect->pad), GST_TIME_ARGS (best_time));
return gst_avi_mux_do_buffer (avimux, best_pad);
} else {
/* simply finish off the file and send EOS */
gst_avi_mux_stop_file (avimux);
gst_pad_push_event (avimux->srcpad, gst_event_new_eos ());
return GST_FLOW_UNEXPECTED;
}
}
static GstFlowReturn
gst_avi_mux_collect_pads (GstCollectPads * pads, GstAviMux * avimux)
{
GstFlowReturn res;
if (G_UNLIKELY (avimux->write_header)) {
if ((res = gst_avi_mux_start_file (avimux)) != GST_FLOW_OK)
return res;
}
return gst_avi_mux_do_one_buffer (avimux);
}
static void
gst_avi_mux_get_property (GObject * object,
guint prop_id, GValue * value, GParamSpec * pspec)
{
GstAviMux *avimux;
avimux = GST_AVI_MUX (object);
switch (prop_id) {
case ARG_BIGFILE:
g_value_set_boolean (value, avimux->enable_large_avi);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static void
gst_avi_mux_set_property (GObject * object,
guint prop_id, const GValue * value, GParamSpec * pspec)
{
GstAviMux *avimux;
avimux = GST_AVI_MUX (object);
switch (prop_id) {
case ARG_BIGFILE:
avimux->enable_large_avi = g_value_get_boolean (value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static GstStateChangeReturn
gst_avi_mux_change_state (GstElement * element, GstStateChange transition)
{
GstAviMux *avimux;
GstStateChangeReturn ret;
avimux = GST_AVI_MUX (element);
switch (transition) {
case GST_STATE_CHANGE_READY_TO_PAUSED:
gst_collect_pads_start (avimux->collect);
break;
case GST_STATE_CHANGE_PAUSED_TO_PLAYING:
break;
case GST_STATE_CHANGE_PAUSED_TO_READY:
gst_collect_pads_stop (avimux->collect);
break;
default:
break;
}
ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
if (ret == GST_STATE_CHANGE_FAILURE)
goto done;
switch (transition) {
case GST_STATE_CHANGE_PLAYING_TO_PAUSED:
break;
case GST_STATE_CHANGE_PAUSED_TO_READY:
gst_avi_mux_reset (avimux);
break;
case GST_STATE_CHANGE_READY_TO_NULL:
break;
default:
break;
}
done:
return ret;
}