/* Quicktime muxer plugin for GStreamer * Copyright (C) 2008-2010 Thiago Santos * Copyright (C) 2008 Mark Nauwelaerts * Copyright (C) 2010 Nokia Corporation. All rights reserved. * Copyright (C) 2014 Jan Schmidt * Contact: Stefan Kost * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /* * Unless otherwise indicated, Source Code is licensed under MIT license. * See further explanation attached in License Statement (distributed in the file * LICENSE). * * Permission is hereby granted, free of charge, to any person obtaining a copy of * this software and associated documentation files (the "Software"), to deal in * the Software without restriction, including without limitation the rights to * use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies * of the Software, and to permit persons to whom the Software is furnished to do * so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ /** * SECTION:element-qtmux * @title: qtmux * @short_description: Muxer for quicktime(.mov) files * * This element merges streams (audio and video) into QuickTime(.mov) files. * * The following background intends to explain why various similar muxers * are present in this plugin. * * The * QuickTime file format specification served as basis for the MP4 file * format specification (mp4mux), and as such the QuickTime file structure is * nearly identical to the so-called ISO Base Media file format defined in * ISO 14496-12 (except for some media specific parts). * In turn, the latter ISO Base Media format was further specialized as a * Motion JPEG-2000 file format in ISO 15444-3 (mj2mux) * and in various 3GPP(2) specs (gppmux). * The fragmented file features defined (only) in ISO Base Media are used by * ISMV files making up (a.o.) Smooth Streaming (ismlmux). * * A few properties (#GstQTMux:movie-timescale, #GstQTMux:trak-timescale, * #GstQTMuxPad:trak-timescale) allow adjusting some technical parameters, * which might be useful in (rare) cases to resolve compatibility issues in * some situations. * * Some other properties influence the result more fundamentally. * A typical mov/mp4 file's metadata (aka moov) is located at the end of the * file, somewhat contrary to this usually being called "the header". * However, a #GstQTMux:faststart file will (with some effort) arrange this to * be located near start of the file, which then allows it e.g. to be played * while downloading. Alternatively, rather than having one chunk of metadata at * start (or end), there can be some metadata at start and most of the other * data can be spread out into fragments of #GstQTMux:fragment-duration. * If such fragmented layout is intended for streaming purposes, then * #GstQTMux:streamable allows foregoing to add index metadata (at the end of * file). * * When the maximum duration to be recorded can be known in advance, #GstQTMux * also supports a 'Robust Muxing' mode. In robust muxing mode, space for the * headers are reserved at the start of muxing, and rewritten at a configurable * interval, so that the output file is always playable, even if the recording * is interrupted uncleanly by a crash. Robust muxing mode requires a seekable * output, such as filesink, because it needs to rewrite the start of the file. * * To enable robust muxing mode, set the #GstQTMux::reserved-moov-update-period * and #GstQTMux::reserved-max-duration property. Also present is the * #GstQTMux::reserved-bytes-per-sec property, which can be increased if * for some reason the default is not large enough and the initial reserved * space for headers is too small. Applications can monitor the * #GstQTMux::reserved-duration-remaining property to see how close to full * the reserved space is becoming. * * Applications that wish to be able to use/edit a file while it is being * written to by live content, can use the "Robust Prefill Muxing" mode. That * mode is a variant of the "Robust Muxing" mode in that it will pre-allocate a * completely valid header from the start for all tracks (i.e. it appears as * though the file is "reserved-max-duration" long with all samples * present). This mode can be enabled by setting the * #GstQTMux::reserved-moov-update-period and #GstQTMux::reserved-prefill * properties. Note that this mode is only possible with input streams that have * a fixed sample size (such as raw audio and Prores Video) and that don't * have reordered samples. * * ## Example pipelines * |[ * gst-launch-1.0 v4l2src num-buffers=500 ! video/x-raw,width=320,height=240 ! videoconvert ! qtmux ! filesink location=video.mov * ]| * Records a video stream captured from a v4l2 device and muxes it into a qt file. * */ /* * Based on avimux */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include #include #include #include #include #include #include #ifdef G_OS_WIN32 #include /* lseek, open, close, read */ #undef lseek #define lseek _lseeki64 #undef off_t #define off_t guint64 #endif #ifdef _MSC_VER #define ftruncate g_win32_ftruncate #endif #ifdef HAVE_UNISTD_H # include #endif #include "gstqtmux.h" GST_DEBUG_CATEGORY_STATIC (gst_qt_mux_debug); #define GST_CAT_DEFAULT gst_qt_mux_debug #ifndef ABSDIFF #define ABSDIFF(a, b) ((a) > (b) ? (a) - (b) : (b) - (a)) #endif /* Hacker notes. * * The basic building blocks of MP4 files are: * - an 'ftyp' box at the very start * - an 'mdat' box which contains the raw audio/video/subtitle data; * this is just a bunch of bytes, completely unframed and possibly * unordered with no additional meta-information * - a 'moov' box that contains information about the different streams * and what they contain, as well as sample tables for each stream * that tell the demuxer where in the mdat box each buffer/sample is * and what its duration/timestamp etc. is, and whether it's a * keyframe etc. * Additionally, fragmented MP4 works by writing chunks of data in * pairs of 'moof' and 'mdat' boxes: * - 'moof' boxes, header preceding each mdat fragment describing the * contents, like a moov but only for that fragment. * - a 'mfra' box for Fragmented MP4, which is written at the end and * contains a summary of all fragments and seek tables. * * Currently mp4mux can work in 4 different modes / generate 4 types * of output files/streams: * * - Normal mp4: mp4mux will write a little ftyp identifier at the * beginning, then start an mdat box into which it will write all the * sample data. At EOS it will then write the moov header with track * headers and sample tables at the end of the file, and rewrite the * start of the file to fix up the mdat box size at the beginning. * It has to wait for EOS to write the moov (which includes the * sample tables) because it doesn't know how much space those * tables will be. The output downstream must be seekable to rewrite * the mdat box at EOS. * * - Fragmented mp4: moov header with track headers at start * but no sample table, followed by N fragments, each containing * track headers with sample tables followed by some data. Downstream * does not need to be seekable if the 'streamable' flag is TRUE, * as the final mfra and total duration will be omitted. * * - Fast-start mp4: the goal here is to create a file where the moov * headers are at the beginning; what mp4mux will do is write all * sample data into a temp file and build moov header plus sample * tables in memory and then when EOS comes, it will push out the * moov header plus sample tables at the beginning, followed by the * mdat sample data at the end which is read in from the temp file * Files created in this mode are better for streaming over the * network, since the client doesn't have to seek to the end of the * file to get the headers, but it requires copying all sample data * out of the temp file at EOS, which can be expensive. Downstream does * not need to be seekable, because of the use of the temp file. * * - Robust Muxing mode: In this mode, qtmux uses the reserved-max-duration * and reserved-moov-update-period properties to reserve free space * at the start of the file and periodically write the MOOV atom out * to it. That means that killing the muxing at any point still * results in a playable file, at the cost of wasting some amount of * free space at the start of file. The approximate recording duration * has to be known in advance to estimate how much free space to reserve * for the moov, and the downstream must be seekable. * If the moov header grows larger than the reserved space, an error * is generated - so it's better to over-estimate the amount of space * to reserve. To ensure the file is playable at any point, the moov * is updated using a 'ping-pong' strategy, so the output is never in * an invalid state. */ #ifndef GST_REMOVE_DEPRECATED enum { DTS_METHOD_DD, DTS_METHOD_REORDER, DTS_METHOD_ASC }; static GType gst_qt_mux_dts_method_get_type (void) { static GType gst_qt_mux_dts_method = 0; if (!gst_qt_mux_dts_method) { static const GEnumValue dts_methods[] = { {DTS_METHOD_DD, "delta/duration", "dd"}, {DTS_METHOD_REORDER, "reorder", "reorder"}, {DTS_METHOD_ASC, "ascending", "asc"}, {0, NULL, NULL}, }; gst_qt_mux_dts_method = g_enum_register_static ("GstQTMuxDtsMethods", dts_methods); } return gst_qt_mux_dts_method; } #define GST_TYPE_QT_MUX_DTS_METHOD \ (gst_qt_mux_dts_method_get_type ()) #endif enum { PROP_PAD_0, PROP_PAD_TRAK_TIMESCALE, }; #define DEFAULT_PAD_TRAK_TIMESCALE 0 GType gst_qt_mux_pad_get_type (void); #define GST_TYPE_QT_MUX_PAD \ (gst_qt_mux_pad_get_type()) #define GST_QT_MUX_PAD(obj) \ (G_TYPE_CHECK_INSTANCE_CAST ((obj), GST_TYPE_QT_MUX_PAD, GstQTMuxPad)) #define GST_QT_MUX_PAD_CLASS(klass) \ (G_TYPE_CHECK_CLASS_CAST ((klass), GST_TYPE_QT_MUX_PAD, GstQTMuxPadClass)) #define GST_IS_QT_MUX_PAD(obj) \ (G_TYPE_CHECK_INSTANCE_TYPE ((obj), GST_TYPE_QT_MUX_PAD)) #define GST_IS_QT_MUX_PAD_CLASS(klass) \ (G_TYPE_CHECK_CLASS_TYPE ((klass), GST_TYPE_QT_MUX_PAD)) #define GST_QT_MUX_PAD_CAST(obj) \ ((GstQTMuxPad *)(obj)) typedef struct _GstQTMuxPad GstQTMuxPad; typedef struct _GstQTMuxPadClass GstQTMuxPadClass; struct _GstQTMuxPad { GstPad parent; guint32 trak_timescale; }; struct _GstQTMuxPadClass { GstPadClass parent; }; G_DEFINE_TYPE (GstQTMuxPad, gst_qt_mux_pad, GST_TYPE_PAD); static void gst_qt_mux_pad_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstQTMuxPad *pad = GST_QT_MUX_PAD_CAST (object); GST_OBJECT_LOCK (pad); switch (prop_id) { case PROP_PAD_TRAK_TIMESCALE: pad->trak_timescale = g_value_get_uint (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } GST_OBJECT_UNLOCK (pad); } static void gst_qt_mux_pad_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstQTMuxPad *pad = GST_QT_MUX_PAD_CAST (object); GST_OBJECT_LOCK (pad); switch (prop_id) { case PROP_PAD_TRAK_TIMESCALE: g_value_set_uint (value, pad->trak_timescale); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } GST_OBJECT_UNLOCK (pad); } static void gst_qt_mux_pad_class_init (GstQTMuxPadClass * klass) { GObjectClass *gobject_class = (GObjectClass *) klass; gobject_class->get_property = gst_qt_mux_pad_get_property; gobject_class->set_property = gst_qt_mux_pad_set_property; g_object_class_install_property (gobject_class, PROP_PAD_TRAK_TIMESCALE, g_param_spec_uint ("trak-timescale", "Track timescale", "Timescale to use for this pad's trak (units per second, 0 is automatic)", 0, G_MAXUINT32, DEFAULT_PAD_TRAK_TIMESCALE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); } static void gst_qt_mux_pad_init (GstQTMuxPad * pad) { pad->trak_timescale = DEFAULT_PAD_TRAK_TIMESCALE; } static guint32 gst_qt_mux_pad_get_timescale (GstQTMuxPad * pad) { guint32 timescale; GST_OBJECT_LOCK (pad); timescale = pad->trak_timescale; GST_OBJECT_UNLOCK (pad); return timescale; } /* QTMux signals and args */ enum { /* FILL ME */ LAST_SIGNAL }; enum { PROP_0, PROP_MOVIE_TIMESCALE, PROP_TRAK_TIMESCALE, PROP_FAST_START, PROP_FAST_START_TEMP_FILE, PROP_MOOV_RECOV_FILE, PROP_FRAGMENT_DURATION, PROP_STREAMABLE, PROP_RESERVED_MAX_DURATION, PROP_RESERVED_DURATION_REMAINING, PROP_RESERVED_MOOV_UPDATE_PERIOD, PROP_RESERVED_BYTES_PER_SEC, PROP_RESERVED_PREFILL, #ifndef GST_REMOVE_DEPRECATED PROP_DTS_METHOD, #endif PROP_DO_CTTS, PROP_INTERLEAVE_BYTES, PROP_INTERLEAVE_TIME, PROP_MAX_RAW_AUDIO_DRIFT, PROP_START_GAP_THRESHOLD, }; /* some spare for header size as well */ #define MDAT_LARGE_FILE_LIMIT ((guint64) 1024 * 1024 * 1024 * 2) #define DEFAULT_MOVIE_TIMESCALE 0 #define DEFAULT_TRAK_TIMESCALE 0 #define DEFAULT_DO_CTTS TRUE #define DEFAULT_FAST_START FALSE #define DEFAULT_FAST_START_TEMP_FILE NULL #define DEFAULT_MOOV_RECOV_FILE NULL #define DEFAULT_FRAGMENT_DURATION 0 #define DEFAULT_STREAMABLE TRUE #ifndef GST_REMOVE_DEPRECATED #define DEFAULT_DTS_METHOD DTS_METHOD_REORDER #endif #define DEFAULT_RESERVED_MAX_DURATION GST_CLOCK_TIME_NONE #define DEFAULT_RESERVED_MOOV_UPDATE_PERIOD GST_CLOCK_TIME_NONE #define DEFAULT_RESERVED_BYTES_PER_SEC_PER_TRAK 550 #define DEFAULT_RESERVED_PREFILL FALSE #define DEFAULT_INTERLEAVE_BYTES 0 #define DEFAULT_INTERLEAVE_TIME 250*GST_MSECOND #define DEFAULT_MAX_RAW_AUDIO_DRIFT 40 * GST_MSECOND #define DEFAULT_START_GAP_THRESHOLD 0 static void gst_qt_mux_finalize (GObject * object); static GstStateChangeReturn gst_qt_mux_change_state (GstElement * element, GstStateChange transition); /* property functions */ static void gst_qt_mux_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_qt_mux_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); /* pad functions */ static GstPad *gst_qt_mux_request_new_pad (GstElement * element, GstPadTemplate * templ, const gchar * name, const GstCaps * caps); static void gst_qt_mux_release_pad (GstElement * element, GstPad * pad); /* event */ static gboolean gst_qt_mux_sink_event (GstCollectPads * pads, GstCollectData * data, GstEvent * event, gpointer user_data); static GstFlowReturn gst_qt_mux_collected (GstCollectPads * pads, gpointer user_data); static GstFlowReturn gst_qt_mux_add_buffer (GstQTMux * qtmux, GstQTPad * pad, GstBuffer * buf); static GstFlowReturn gst_qt_mux_robust_recording_rewrite_moov (GstQTMux * qtmux); static void gst_qt_mux_update_global_statistics (GstQTMux * qtmux); static void gst_qt_mux_update_edit_lists (GstQTMux * qtmux); static GstElementClass *parent_class = NULL; static void gst_qt_mux_base_init (gpointer g_class) { GstElementClass *element_class = GST_ELEMENT_CLASS (g_class); GstQTMuxClass *klass = (GstQTMuxClass *) g_class; GstQTMuxClassParams *params; GstPadTemplate *videosinktempl, *audiosinktempl, *subtitlesinktempl, *captionsinktempl; GstPadTemplate *srctempl; gchar *longname, *description; params = (GstQTMuxClassParams *) g_type_get_qdata (G_OBJECT_CLASS_TYPE (g_class), GST_QT_MUX_PARAMS_QDATA); g_assert (params != NULL); /* construct the element details struct */ longname = g_strdup_printf ("%s Muxer", params->prop->long_name); description = g_strdup_printf ("Multiplex audio and video into a %s file", params->prop->long_name); gst_element_class_set_static_metadata (element_class, longname, "Codec/Muxer", description, "Thiago Sousa Santos "); g_free (longname); g_free (description); /* pad templates */ srctempl = gst_pad_template_new ("src", GST_PAD_SRC, GST_PAD_ALWAYS, params->src_caps); gst_element_class_add_pad_template (element_class, srctempl); if (params->audio_sink_caps) { audiosinktempl = gst_pad_template_new_with_gtype ("audio_%u", GST_PAD_SINK, GST_PAD_REQUEST, params->audio_sink_caps, GST_TYPE_QT_MUX_PAD); gst_element_class_add_pad_template (element_class, audiosinktempl); } if (params->video_sink_caps) { videosinktempl = gst_pad_template_new_with_gtype ("video_%u", GST_PAD_SINK, GST_PAD_REQUEST, params->video_sink_caps, GST_TYPE_QT_MUX_PAD); gst_element_class_add_pad_template (element_class, videosinktempl); } if (params->subtitle_sink_caps) { subtitlesinktempl = gst_pad_template_new_with_gtype ("subtitle_%u", GST_PAD_SINK, GST_PAD_REQUEST, params->subtitle_sink_caps, GST_TYPE_QT_MUX_PAD); gst_element_class_add_pad_template (element_class, subtitlesinktempl); } if (params->caption_sink_caps) { captionsinktempl = gst_pad_template_new_with_gtype ("caption_%u", GST_PAD_SINK, GST_PAD_REQUEST, params->caption_sink_caps, GST_TYPE_QT_MUX_PAD); gst_element_class_add_pad_template (element_class, captionsinktempl); } klass->format = params->prop->format; } static void gst_qt_mux_class_init (GstQTMuxClass * klass) { GObjectClass *gobject_class; GstElementClass *gstelement_class; GParamFlags streamable_flags; const gchar *streamable_desc; gboolean streamable; #define STREAMABLE_DESC "If set to true, the output should be as if it is to "\ "be streamed and hence no indexes written or duration written." gobject_class = (GObjectClass *) klass; gstelement_class = (GstElementClass *) klass; parent_class = g_type_class_peek_parent (klass); gobject_class->finalize = gst_qt_mux_finalize; gobject_class->get_property = gst_qt_mux_get_property; gobject_class->set_property = gst_qt_mux_set_property; streamable_flags = G_PARAM_READWRITE | G_PARAM_CONSTRUCT; if (klass->format == GST_QT_MUX_FORMAT_ISML) { streamable_desc = STREAMABLE_DESC; streamable = DEFAULT_STREAMABLE; } else { streamable_desc = STREAMABLE_DESC " (DEPRECATED, only valid for fragmented MP4)"; streamable_flags |= G_PARAM_DEPRECATED; streamable = FALSE; } g_object_class_install_property (gobject_class, PROP_MOVIE_TIMESCALE, g_param_spec_uint ("movie-timescale", "Movie timescale", "Timescale to use in the movie (units per second, 0 == default)", 0, G_MAXUINT32, DEFAULT_MOVIE_TIMESCALE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_TRAK_TIMESCALE, g_param_spec_uint ("trak-timescale", "Track timescale", "Timescale to use for the tracks (units per second, 0 is automatic)", 0, G_MAXUINT32, DEFAULT_TRAK_TIMESCALE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_DO_CTTS, g_param_spec_boolean ("presentation-time", "Include presentation-time info", "Calculate and include presentation/composition time " "(in addition to decoding time)", DEFAULT_DO_CTTS, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); #ifndef GST_REMOVE_DEPRECATED g_object_class_install_property (gobject_class, PROP_DTS_METHOD, g_param_spec_enum ("dts-method", "dts-method", "Method to determine DTS time (DEPRECATED)", GST_TYPE_QT_MUX_DTS_METHOD, DEFAULT_DTS_METHOD, G_PARAM_DEPRECATED | G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); #endif g_object_class_install_property (gobject_class, PROP_FAST_START, g_param_spec_boolean ("faststart", "Format file to faststart", "If the file should be formatted for faststart (headers first)", DEFAULT_FAST_START, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_FAST_START_TEMP_FILE, g_param_spec_string ("faststart-file", "File to use for storing buffers", "File that will be used temporarily to store data from the stream " "when creating a faststart file. If null a filepath will be " "created automatically", DEFAULT_FAST_START_TEMP_FILE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS | GST_PARAM_DOC_SHOW_DEFAULT)); g_object_class_install_property (gobject_class, PROP_MOOV_RECOV_FILE, g_param_spec_string ("moov-recovery-file", "File to store data for posterior moov atom recovery", "File to be used to store " "data for moov atom making movie file recovery possible in case " "of a crash during muxing. Null for disabled. (Experimental)", DEFAULT_MOOV_RECOV_FILE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_FRAGMENT_DURATION, g_param_spec_uint ("fragment-duration", "Fragment duration", "Fragment durations in ms (produce a fragmented file if > 0)", 0, G_MAXUINT32, klass->format == GST_QT_MUX_FORMAT_ISML ? 2000 : DEFAULT_FRAGMENT_DURATION, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_STREAMABLE, g_param_spec_boolean ("streamable", "Streamable", streamable_desc, streamable, streamable_flags | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_RESERVED_MAX_DURATION, g_param_spec_uint64 ("reserved-max-duration", "Reserved maximum file duration (ns)", "When set to a value > 0, reserves space for index tables at the " "beginning of the file.", 0, G_MAXUINT64, DEFAULT_RESERVED_MAX_DURATION, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_RESERVED_DURATION_REMAINING, g_param_spec_uint64 ("reserved-duration-remaining", "Report the approximate amount of remaining recording space (ns)", "Reports the approximate amount of remaining moov header space " "reserved using reserved-max-duration", 0, G_MAXUINT64, 0, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_RESERVED_MOOV_UPDATE_PERIOD, g_param_spec_uint64 ("reserved-moov-update-period", "Interval at which to update index tables (ns)", "When used with reserved-max-duration, periodically updates the " "index tables with information muxed so far.", 0, G_MAXUINT64, DEFAULT_RESERVED_MOOV_UPDATE_PERIOD, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_RESERVED_BYTES_PER_SEC, g_param_spec_uint ("reserved-bytes-per-sec", "Reserved MOOV bytes per second, per track", "Multiplier for converting reserved-max-duration into bytes of header to reserve, per second, per track", 0, 10000, DEFAULT_RESERVED_BYTES_PER_SEC_PER_TRAK, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_RESERVED_PREFILL, g_param_spec_boolean ("reserved-prefill", "Reserved Prefill Samples Table", "Prefill samples table of reserved duration", DEFAULT_RESERVED_PREFILL, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_INTERLEAVE_BYTES, g_param_spec_uint64 ("interleave-bytes", "Interleave (bytes)", "Interleave between streams in bytes", 0, G_MAXUINT64, DEFAULT_INTERLEAVE_BYTES, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_INTERLEAVE_TIME, g_param_spec_uint64 ("interleave-time", "Interleave (time)", "Interleave between streams in nanoseconds", 0, G_MAXUINT64, DEFAULT_INTERLEAVE_TIME, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_MAX_RAW_AUDIO_DRIFT, g_param_spec_uint64 ("max-raw-audio-drift", "Max Raw Audio Drift", "Maximum allowed drift of raw audio samples vs. timestamps in nanoseconds", 0, G_MAXUINT64, DEFAULT_MAX_RAW_AUDIO_DRIFT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_START_GAP_THRESHOLD, g_param_spec_uint64 ("start-gap-threshold", "Start Gap Threshold", "Threshold for creating an edit list for gaps at the start in nanoseconds", 0, G_MAXUINT64, DEFAULT_START_GAP_THRESHOLD, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); gstelement_class->request_new_pad = GST_DEBUG_FUNCPTR (gst_qt_mux_request_new_pad); gstelement_class->change_state = GST_DEBUG_FUNCPTR (gst_qt_mux_change_state); gstelement_class->release_pad = GST_DEBUG_FUNCPTR (gst_qt_mux_release_pad); } static void gst_qt_mux_pad_reset (GstQTPad * qtpad) { qtpad->fourcc = 0; qtpad->is_out_of_order = FALSE; qtpad->sample_size = 0; qtpad->sync = FALSE; qtpad->last_dts = 0; qtpad->sample_offset = 0; qtpad->dts_adjustment = GST_CLOCK_TIME_NONE; qtpad->first_ts = GST_CLOCK_TIME_NONE; qtpad->first_dts = GST_CLOCK_TIME_NONE; qtpad->prepare_buf_func = NULL; qtpad->create_empty_buffer = NULL; qtpad->avg_bitrate = 0; qtpad->max_bitrate = 0; qtpad->total_duration = 0; qtpad->total_bytes = 0; qtpad->sparse = FALSE; gst_buffer_replace (&qtpad->last_buf, NULL); if (qtpad->tags) { gst_tag_list_unref (qtpad->tags); qtpad->tags = NULL; } /* reference owned elsewhere */ qtpad->trak = NULL; qtpad->tc_trak = NULL; if (qtpad->traf) { atom_traf_free (qtpad->traf); qtpad->traf = NULL; } atom_array_clear (&qtpad->fragment_buffers); if (qtpad->samples) g_array_unref (qtpad->samples); qtpad->samples = NULL; /* reference owned elsewhere */ qtpad->tfra = NULL; qtpad->first_pts = GST_CLOCK_TIME_NONE; qtpad->tc_pos = -1; if (qtpad->first_tc) gst_video_time_code_free (qtpad->first_tc); qtpad->first_tc = NULL; if (qtpad->raw_audio_adapter) gst_object_unref (qtpad->raw_audio_adapter); qtpad->raw_audio_adapter = NULL; } /* * Takes GstQTMux back to its initial state */ static void gst_qt_mux_reset (GstQTMux * qtmux, gboolean alloc) { GSList *walk; qtmux->state = GST_QT_MUX_STATE_NONE; qtmux->header_size = 0; qtmux->mdat_size = 0; qtmux->moov_pos = 0; qtmux->mdat_pos = 0; qtmux->longest_chunk = GST_CLOCK_TIME_NONE; qtmux->fragment_sequence = 0; if (qtmux->ftyp) { atom_ftyp_free (qtmux->ftyp); qtmux->ftyp = NULL; } if (qtmux->moov) { atom_moov_free (qtmux->moov); qtmux->moov = NULL; } if (qtmux->mfra) { atom_mfra_free (qtmux->mfra); qtmux->mfra = NULL; } if (qtmux->fast_start_file) { fclose (qtmux->fast_start_file); g_remove (qtmux->fast_start_file_path); qtmux->fast_start_file = NULL; } if (qtmux->moov_recov_file) { fclose (qtmux->moov_recov_file); qtmux->moov_recov_file = NULL; } for (walk = qtmux->extra_atoms; walk; walk = g_slist_next (walk)) { AtomInfo *ainfo = (AtomInfo *) walk->data; ainfo->free_func (ainfo->atom); g_free (ainfo); } g_slist_free (qtmux->extra_atoms); qtmux->extra_atoms = NULL; GST_OBJECT_LOCK (qtmux); gst_tag_setter_reset_tags (GST_TAG_SETTER (qtmux)); GST_OBJECT_UNLOCK (qtmux); /* reset pad data */ for (walk = qtmux->sinkpads; walk; walk = g_slist_next (walk)) { GstQTPad *qtpad = (GstQTPad *) walk->data; gst_qt_mux_pad_reset (qtpad); /* hm, moov_free above yanked the traks away from us, * so do not free, but do clear */ qtpad->trak = NULL; } if (alloc) { qtmux->moov = atom_moov_new (qtmux->context); /* ensure all is as nice and fresh as request_new_pad would provide it */ for (walk = qtmux->sinkpads; walk; walk = g_slist_next (walk)) { GstQTPad *qtpad = (GstQTPad *) walk->data; qtpad->trak = atom_trak_new (qtmux->context); atom_moov_add_trak (qtmux->moov, qtpad->trak); } } qtmux->current_pad = NULL; qtmux->current_chunk_size = 0; qtmux->current_chunk_duration = 0; qtmux->current_chunk_offset = -1; qtmux->reserved_moov_size = 0; qtmux->last_moov_update = GST_CLOCK_TIME_NONE; qtmux->muxed_since_last_update = 0; qtmux->reserved_duration_remaining = GST_CLOCK_TIME_NONE; } static void gst_qt_mux_init (GstQTMux * qtmux, GstQTMuxClass * qtmux_klass) { GstElementClass *klass = GST_ELEMENT_CLASS (qtmux_klass); GstPadTemplate *templ; templ = gst_element_class_get_pad_template (klass, "src"); qtmux->srcpad = gst_pad_new_from_template (templ, "src"); gst_pad_use_fixed_caps (qtmux->srcpad); gst_element_add_pad (GST_ELEMENT (qtmux), qtmux->srcpad); qtmux->sinkpads = NULL; qtmux->collect = gst_collect_pads_new (); gst_collect_pads_set_event_function (qtmux->collect, GST_DEBUG_FUNCPTR (gst_qt_mux_sink_event), qtmux); gst_collect_pads_set_clip_function (qtmux->collect, GST_DEBUG_FUNCPTR (gst_collect_pads_clip_running_time), qtmux); gst_collect_pads_set_function (qtmux->collect, GST_DEBUG_FUNCPTR (gst_qt_mux_collected), qtmux); /* properties set to default upon construction */ qtmux->reserved_max_duration = DEFAULT_RESERVED_MAX_DURATION; qtmux->reserved_moov_update_period = DEFAULT_RESERVED_MOOV_UPDATE_PERIOD; qtmux->reserved_bytes_per_sec_per_trak = DEFAULT_RESERVED_BYTES_PER_SEC_PER_TRAK; qtmux->interleave_bytes = DEFAULT_INTERLEAVE_BYTES; qtmux->interleave_time = DEFAULT_INTERLEAVE_TIME; qtmux->max_raw_audio_drift = DEFAULT_MAX_RAW_AUDIO_DRIFT; qtmux->start_gap_threshold = DEFAULT_START_GAP_THRESHOLD; /* always need this */ qtmux->context = atoms_context_new (gst_qt_mux_map_format_to_flavor (qtmux_klass->format)); /* internals to initial state */ gst_qt_mux_reset (qtmux, TRUE); } static void gst_qt_mux_finalize (GObject * object) { GstQTMux *qtmux = GST_QT_MUX_CAST (object); gst_qt_mux_reset (qtmux, FALSE); g_free (qtmux->fast_start_file_path); g_free (qtmux->moov_recov_file_path); atoms_context_free (qtmux->context); gst_object_unref (qtmux->collect); g_slist_free (qtmux->sinkpads); G_OBJECT_CLASS (parent_class)->finalize (object); } static GstBuffer * gst_qt_mux_prepare_jpc_buffer (GstQTPad * qtpad, GstBuffer * buf, GstQTMux * qtmux) { GstBuffer *newbuf; GstMapInfo map; gsize size; GST_LOG_OBJECT (qtmux, "Preparing jpc buffer"); if (buf == NULL) return NULL; size = gst_buffer_get_size (buf); newbuf = gst_buffer_new_and_alloc (size + 8); gst_buffer_copy_into (newbuf, buf, GST_BUFFER_COPY_ALL, 8, size); gst_buffer_map (newbuf, &map, GST_MAP_WRITE); GST_WRITE_UINT32_BE (map.data, map.size); GST_WRITE_UINT32_LE (map.data + 4, FOURCC_jp2c); gst_buffer_unmap (buf, &map); gst_buffer_unref (buf); return newbuf; } static gsize extract_608_field_from_s334_1a (const guint8 * ccdata, gsize ccdata_size, guint field, guint8 ** res) { guint8 *storage; gsize storage_size = 128; gsize i, res_size = 0; storage = g_malloc0 (storage_size); /* Iterate over the ccdata and put the corresponding tuples for the given field * in the storage */ for (i = 0; i < ccdata_size; i += 3) { if ((field == 1 && (ccdata[i * 3] & 0x80)) || (field == 2 && !(ccdata[i * 3] & 0x80))) { GST_DEBUG ("Storing matching cc for field %d : 0x%02x 0x%02x", field, ccdata[i * 3 + 1], ccdata[i * 3 + 2]); if (res_size >= storage_size) { storage_size += 128; storage = g_realloc (storage, storage_size); } storage[res_size] = ccdata[i * 3 + 1]; storage[res_size + 1] = ccdata[i * 3 + 2]; res_size += 2; } } if (res_size == 0) { g_free (storage); *res = NULL; return 0; } *res = storage; return res_size; } static GstBuffer * gst_qt_mux_prepare_caption_buffer (GstQTPad * qtpad, GstBuffer * buf, GstQTMux * qtmux) { GstBuffer *newbuf = NULL; GstMapInfo map, inmap; gsize size; gboolean in_prefill; if (buf == NULL) return NULL; in_prefill = (qtmux->mux_mode == GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL); size = gst_buffer_get_size (buf); gst_buffer_map (buf, &inmap, GST_MAP_READ); GST_LOG_OBJECT (qtmux, "Preparing caption buffer %" GST_FOURCC_FORMAT " size:%" G_GSIZE_FORMAT, GST_FOURCC_ARGS (qtpad->fourcc), size); switch (qtpad->fourcc) { case FOURCC_c608: { guint8 *cdat, *cdt2; gsize cdat_size, cdt2_size, total_size = 0; gsize write_offs = 0; cdat_size = extract_608_field_from_s334_1a (inmap.data, inmap.size, 1, &cdat); cdt2_size = extract_608_field_from_s334_1a (inmap.data, inmap.size, 2, &cdt2); if (cdat_size) total_size += cdat_size + 8; if (cdt2_size) total_size += cdt2_size + 8; if (total_size == 0) { GST_DEBUG_OBJECT (qtmux, "No 608 data ?"); /* FIXME : We might want to *always* store something, even if * it's "empty" CC (i.e. 0x80 0x80) */ break; } newbuf = gst_buffer_new_and_alloc (in_prefill ? 20 : total_size); /* Let's copy over all metadata and not the memory */ gst_buffer_copy_into (newbuf, buf, GST_BUFFER_COPY_METADATA, 0, size); gst_buffer_map (newbuf, &map, GST_MAP_WRITE); if (cdat_size || in_prefill) { GST_WRITE_UINT32_BE (map.data, in_prefill ? 10 : cdat_size + 8); GST_WRITE_UINT32_LE (map.data + 4, FOURCC_cdat); if (cdat_size) memcpy (map.data + 8, cdat, in_prefill ? 2 : cdat_size); else { /* Write 'empty' CC */ map.data[8] = 0x80; map.data[9] = 0x80; } write_offs = in_prefill ? 10 : cdat_size + 8; if (cdat_size) g_free (cdat); } if (cdt2_size || in_prefill) { GST_WRITE_UINT32_BE (map.data + write_offs, in_prefill ? 10 : cdt2_size + 8); GST_WRITE_UINT32_LE (map.data + write_offs + 4, FOURCC_cdt2); if (cdt2_size) memcpy (map.data + write_offs + 8, cdt2, in_prefill ? 2 : cdt2_size); else { /* Write 'empty' CC */ map.data[write_offs + 8] = 0x80; map.data[write_offs + 9] = 0x80; } if (cdt2_size) g_free (cdt2); } gst_buffer_unmap (newbuf, &map); break; } break; case FOURCC_c708: { /* Take the whole CDP */ if (in_prefill && size > 256) { GST_ERROR_OBJECT (qtmux, "Input C708 CDP too big for prefill mode !"); break; } newbuf = gst_buffer_new_and_alloc (in_prefill ? 256 + 8 : size + 8); /* Let's copy over all metadata and not the memory */ gst_buffer_copy_into (newbuf, buf, GST_BUFFER_COPY_METADATA, 0, size); gst_buffer_map (newbuf, &map, GST_MAP_WRITE); GST_WRITE_UINT32_BE (map.data, size + 8); GST_WRITE_UINT32_LE (map.data + 4, FOURCC_ccdp); memcpy (map.data + 8, inmap.data, inmap.size); gst_buffer_unmap (newbuf, &map); break; } default: /* theoretically this should never happen, but let's keep this here in case */ GST_WARNING_OBJECT (qtmux, "Unknown caption format"); break; } gst_buffer_unmap (buf, &inmap); gst_buffer_unref (buf); return newbuf; } static GstBuffer * gst_qt_mux_prepare_tx3g_buffer (GstQTPad * qtpad, GstBuffer * buf, GstQTMux * qtmux) { GstBuffer *newbuf; GstMapInfo frommap; GstMapInfo tomap; gsize size; const guint8 *dataend; GST_LOG_OBJECT (qtmux, "Preparing tx3g buffer %" GST_PTR_FORMAT, buf); if (buf == NULL) return NULL; gst_buffer_map (buf, &frommap, GST_MAP_READ); dataend = memchr (frommap.data, 0, frommap.size); size = dataend ? dataend - frommap.data : frommap.size; newbuf = gst_buffer_new_and_alloc (size + 2); gst_buffer_map (newbuf, &tomap, GST_MAP_WRITE); GST_WRITE_UINT16_BE (tomap.data, size); memcpy (tomap.data + 2, frommap.data, size); gst_buffer_unmap (newbuf, &tomap); gst_buffer_unmap (buf, &frommap); gst_buffer_copy_into (newbuf, buf, GST_BUFFER_COPY_METADATA, 0, size); /* gst_buffer_copy_into is trying to be too clever and * won't copy duration when size is different */ GST_BUFFER_DURATION (newbuf) = GST_BUFFER_DURATION (buf); gst_buffer_unref (buf); return newbuf; } static void gst_qt_mux_pad_add_ac3_extension (GstQTMux * qtmux, GstQTPad * qtpad, guint8 fscod, guint8 frmsizcod, guint8 bsid, guint8 bsmod, guint8 acmod, guint8 lfe_on) { AtomInfo *ext; g_return_if_fail (qtpad->trak_ste); ext = build_ac3_extension (fscod, bsid, bsmod, acmod, lfe_on, frmsizcod >> 1); /* bitrate_code is inside frmsizcod */ sample_table_entry_add_ext_atom (qtpad->trak_ste, ext); } static GstBuffer * gst_qt_mux_prepare_parse_ac3_frame (GstQTPad * qtpad, GstBuffer * buf, GstQTMux * qtmux) { GstMapInfo map; GstByteReader reader; guint off; if (!gst_buffer_map (buf, &map, GST_MAP_READ)) { GST_WARNING_OBJECT (qtpad->collect.pad, "Failed to map buffer"); return buf; } if (G_UNLIKELY (map.size < 8)) goto done; gst_byte_reader_init (&reader, map.data, map.size); off = gst_byte_reader_masked_scan_uint32 (&reader, 0xffff0000, 0x0b770000, 0, map.size); if (off != -1) { GstBitReader bits; guint8 fscod, frmsizcod, bsid, bsmod, acmod, lfe_on; GST_DEBUG_OBJECT (qtpad->collect.pad, "Found ac3 sync point at offset: %u", off); gst_bit_reader_init (&bits, map.data, map.size); /* off + sync + crc */ gst_bit_reader_skip_unchecked (&bits, off * 8 + 16 + 16); fscod = gst_bit_reader_get_bits_uint8_unchecked (&bits, 2); frmsizcod = gst_bit_reader_get_bits_uint8_unchecked (&bits, 6); bsid = gst_bit_reader_get_bits_uint8_unchecked (&bits, 5); bsmod = gst_bit_reader_get_bits_uint8_unchecked (&bits, 3); acmod = gst_bit_reader_get_bits_uint8_unchecked (&bits, 3); if ((acmod & 0x1) && (acmod != 0x1)) /* 3 front channels */ gst_bit_reader_skip_unchecked (&bits, 2); if ((acmod & 0x4)) /* if a surround channel exists */ gst_bit_reader_skip_unchecked (&bits, 2); if (acmod == 0x2) /* if in 2/0 mode */ gst_bit_reader_skip_unchecked (&bits, 2); lfe_on = gst_bit_reader_get_bits_uint8_unchecked (&bits, 1); gst_qt_mux_pad_add_ac3_extension (qtmux, qtpad, fscod, frmsizcod, bsid, bsmod, acmod, lfe_on); /* AC-3 spec says that those values should be constant for the * whole stream when muxed in mp4. We trust the input follows it */ GST_DEBUG_OBJECT (qtpad->collect.pad, "Data parsed, removing " "prepare buffer function"); qtpad->prepare_buf_func = NULL; } done: gst_buffer_unmap (buf, &map); return buf; } static GstBuffer * gst_qt_mux_create_empty_tx3g_buffer (GstQTPad * qtpad, gint64 duration) { guint8 *data; data = g_malloc (2); GST_WRITE_UINT16_BE (data, 0); return gst_buffer_new_wrapped (data, 2); } static void gst_qt_mux_add_mp4_tag (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc) { switch (gst_tag_get_type (tag)) { /* strings */ case G_TYPE_STRING: { gchar *str = NULL; if (!gst_tag_list_get_string (list, tag, &str) || !str) break; GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %s", GST_FOURCC_ARGS (fourcc), str); atom_udta_add_str_tag (udta, fourcc, str); g_free (str); break; } /* double */ case G_TYPE_DOUBLE: { gdouble value; if (!gst_tag_list_get_double (list, tag, &value)) break; GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %u", GST_FOURCC_ARGS (fourcc), (gint) value); atom_udta_add_uint_tag (udta, fourcc, 21, (gint) value); break; } case G_TYPE_UINT: { guint value = 0; if (tag2) { /* paired unsigned integers */ guint count = 0; gboolean got_tag; got_tag = gst_tag_list_get_uint (list, tag, &value); got_tag = gst_tag_list_get_uint (list, tag2, &count) || got_tag; if (!got_tag) break; GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %u/%u", GST_FOURCC_ARGS (fourcc), value, count); atom_udta_add_uint_tag (udta, fourcc, 0, value << 16 | (count & 0xFFFF)); } else { /* unpaired unsigned integers */ if (!gst_tag_list_get_uint (list, tag, &value)) break; GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %u", GST_FOURCC_ARGS (fourcc), value); atom_udta_add_uint_tag (udta, fourcc, 1, value); } break; } default: g_assert_not_reached (); break; } } static void gst_qt_mux_add_mp4_date (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc) { GDate *date = NULL; GDateYear year; GDateMonth month; GDateDay day; gchar *str; g_return_if_fail (gst_tag_get_type (tag) == G_TYPE_DATE); if (!gst_tag_list_get_date (list, tag, &date) || !date) return; year = g_date_get_year (date); month = g_date_get_month (date); day = g_date_get_day (date); g_date_free (date); if (year == G_DATE_BAD_YEAR && month == G_DATE_BAD_MONTH && day == G_DATE_BAD_DAY) { GST_WARNING_OBJECT (qtmux, "invalid date in tag"); return; } str = g_strdup_printf ("%u-%u-%u", year, month, day); GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %s", GST_FOURCC_ARGS (fourcc), str); atom_udta_add_str_tag (udta, fourcc, str); g_free (str); } static void gst_qt_mux_add_mp4_cover (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc) { GValue value = { 0, }; GstBuffer *buf; GstSample *sample; GstCaps *caps; GstStructure *structure; gint flags = 0; GstMapInfo map; g_return_if_fail (gst_tag_get_type (tag) == GST_TYPE_SAMPLE); if (!gst_tag_list_copy_value (&value, list, tag)) return; sample = gst_value_get_sample (&value); if (!sample) goto done; buf = gst_sample_get_buffer (sample); if (!buf) goto done; caps = gst_sample_get_caps (sample); if (!caps) { GST_WARNING_OBJECT (qtmux, "preview image without caps"); goto done; } GST_DEBUG_OBJECT (qtmux, "preview image caps %" GST_PTR_FORMAT, caps); structure = gst_caps_get_structure (caps, 0); if (gst_structure_has_name (structure, "image/jpeg")) flags = 13; else if (gst_structure_has_name (structure, "image/png")) flags = 14; if (!flags) { GST_WARNING_OBJECT (qtmux, "preview image format not supported"); goto done; } gst_buffer_map (buf, &map, GST_MAP_READ); GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> image size %" G_GSIZE_FORMAT "", GST_FOURCC_ARGS (fourcc), map.size); atom_udta_add_tag (udta, fourcc, flags, map.data, map.size); gst_buffer_unmap (buf, &map); done: g_value_unset (&value); } static void gst_qt_mux_add_3gp_str (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc) { gchar *str = NULL; guint number; g_return_if_fail (gst_tag_get_type (tag) == G_TYPE_STRING); g_return_if_fail (!tag2 || gst_tag_get_type (tag2) == G_TYPE_UINT); if (!gst_tag_list_get_string (list, tag, &str) || !str) return; if (tag2) if (!gst_tag_list_get_uint (list, tag2, &number)) tag2 = NULL; if (!tag2) { GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %s", GST_FOURCC_ARGS (fourcc), str); atom_udta_add_3gp_str_tag (udta, fourcc, str); } else { GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %s/%d", GST_FOURCC_ARGS (fourcc), str, number); atom_udta_add_3gp_str_int_tag (udta, fourcc, str, number); } g_free (str); } static void gst_qt_mux_add_3gp_date (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc) { GDate *date = NULL; GDateYear year; g_return_if_fail (gst_tag_get_type (tag) == G_TYPE_DATE); if (!gst_tag_list_get_date (list, tag, &date) || !date) return; year = g_date_get_year (date); g_date_free (date); if (year == G_DATE_BAD_YEAR) { GST_WARNING_OBJECT (qtmux, "invalid date in tag"); return; } GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %d", GST_FOURCC_ARGS (fourcc), year); atom_udta_add_3gp_uint_tag (udta, fourcc, year); } static void gst_qt_mux_add_3gp_location (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc) { gdouble latitude = -360, longitude = -360, altitude = 0; gchar *location = NULL; guint8 *data, *ddata; gint size = 0, len = 0; gboolean ret = FALSE; g_return_if_fail (strcmp (tag, GST_TAG_GEO_LOCATION_NAME) == 0); ret = gst_tag_list_get_string (list, tag, &location); ret |= gst_tag_list_get_double (list, GST_TAG_GEO_LOCATION_LONGITUDE, &longitude); ret |= gst_tag_list_get_double (list, GST_TAG_GEO_LOCATION_LATITUDE, &latitude); ret |= gst_tag_list_get_double (list, GST_TAG_GEO_LOCATION_ELEVATION, &altitude); if (!ret) return; if (location) len = strlen (location); size += len + 1 + 2; /* role + (long, lat, alt) + body + notes */ size += 1 + 3 * 4 + 1 + 1; data = ddata = g_malloc (size); /* language tag */ GST_WRITE_UINT16_BE (data, language_code (GST_QT_MUX_DEFAULT_TAG_LANGUAGE)); /* location */ if (location) memcpy (data + 2, location, len); GST_WRITE_UINT8 (data + 2 + len, 0); data += len + 1 + 2; /* role */ GST_WRITE_UINT8 (data, 0); /* long, lat, alt */ #define QT_WRITE_SFP32(data, fp) GST_WRITE_UINT32_BE(data, (guint32) ((gint) (fp * 65536.0))) QT_WRITE_SFP32 (data + 1, longitude); QT_WRITE_SFP32 (data + 5, latitude); QT_WRITE_SFP32 (data + 9, altitude); /* neither astronomical body nor notes */ GST_WRITE_UINT16_BE (data + 13, 0); GST_DEBUG_OBJECT (qtmux, "Adding tag 'loci'"); atom_udta_add_3gp_tag (udta, fourcc, ddata, size); g_free (ddata); } static void gst_qt_mux_add_3gp_keywords (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc) { gchar *keywords = NULL; guint8 *data, *ddata; gint size = 0, i; gchar **kwds; g_return_if_fail (strcmp (tag, GST_TAG_KEYWORDS) == 0); if (!gst_tag_list_get_string (list, tag, &keywords) || !keywords) return; kwds = g_strsplit (keywords, ",", 0); g_free (keywords); size = 0; for (i = 0; kwds[i]; i++) { /* size byte + null-terminator */ size += strlen (kwds[i]) + 1 + 1; } /* language tag + count + keywords */ size += 2 + 1; data = ddata = g_malloc (size); /* language tag */ GST_WRITE_UINT16_BE (data, language_code (GST_QT_MUX_DEFAULT_TAG_LANGUAGE)); /* count */ GST_WRITE_UINT8 (data + 2, i); data += 3; /* keywords */ for (i = 0; kwds[i]; ++i) { gint len = strlen (kwds[i]); GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %s", GST_FOURCC_ARGS (fourcc), kwds[i]); /* size */ GST_WRITE_UINT8 (data, len + 1); memcpy (data + 1, kwds[i], len + 1); data += len + 2; } g_strfreev (kwds); atom_udta_add_3gp_tag (udta, fourcc, ddata, size); g_free (ddata); } static gboolean gst_qt_mux_parse_classification_string (GstQTMux * qtmux, const gchar * input, guint32 * p_fourcc, guint16 * p_table, gchar ** p_content) { guint32 fourcc; gint table; gint size; const gchar *data; data = input; size = strlen (input); if (size < 4 + 3 + 1 + 1 + 1) { /* at least the minimum xxxx://y/z */ GST_WARNING_OBJECT (qtmux, "Classification tag input (%s) too short, " "ignoring", input); return FALSE; } /* read the fourcc */ memcpy (&fourcc, data, 4); size -= 4; data += 4; if (strncmp (data, "://", 3) != 0) { goto mismatch; } data += 3; size -= 3; /* read the table number */ if (sscanf (data, "%d", &table) != 1) { goto mismatch; } if (table < 0) { GST_WARNING_OBJECT (qtmux, "Invalid table number in classification tag (%d)" ", table numbers should be positive, ignoring tag", table); return FALSE; } /* find the next / */ while (size > 0 && data[0] != '/') { data += 1; size -= 1; } if (size == 0) { goto mismatch; } g_assert (data[0] == '/'); /* skip the '/' */ data += 1; size -= 1; if (size == 0) { goto mismatch; } /* read up the rest of the string */ *p_content = g_strdup (data); *p_table = (guint16) table; *p_fourcc = fourcc; return TRUE; mismatch: { GST_WARNING_OBJECT (qtmux, "Ignoring classification tag as " "input (%s) didn't match the expected entitycode://table/content", input); return FALSE; } } static void gst_qt_mux_add_3gp_classification (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc) { gchar *clsf_data = NULL; gint size = 0; guint32 entity = 0; guint16 table = 0; gchar *content = NULL; guint8 *data; g_return_if_fail (strcmp (tag, GST_TAG_3GP_CLASSIFICATION) == 0); if (!gst_tag_list_get_string (list, tag, &clsf_data) || !clsf_data) return; GST_DEBUG_OBJECT (qtmux, "Adding tag %" GST_FOURCC_FORMAT " -> %s", GST_FOURCC_ARGS (fourcc), clsf_data); /* parse the string, format is: * entityfourcc://table/content */ gst_qt_mux_parse_classification_string (qtmux, clsf_data, &entity, &table, &content); g_free (clsf_data); /* +1 for the \0 */ size = strlen (content) + 1; /* now we have everything, build the atom * atom description is at 3GPP TS 26.244 V8.2.0 (2009-09) */ data = g_malloc (4 + 2 + 2 + size); GST_WRITE_UINT32_LE (data, entity); GST_WRITE_UINT16_BE (data + 4, (guint16) table); GST_WRITE_UINT16_BE (data + 6, 0); memcpy (data + 8, content, size); g_free (content); atom_udta_add_3gp_tag (udta, fourcc, data, 4 + 2 + 2 + size); g_free (data); } typedef void (*GstQTMuxAddUdtaTagFunc) (GstQTMux * mux, const GstTagList * list, AtomUDTA * udta, const char *tag, const char *tag2, guint32 fourcc); /* * Struct to record mappings from gstreamer tags to fourcc codes */ typedef struct _GstTagToFourcc { guint32 fourcc; const gchar *gsttag; const gchar *gsttag2; const GstQTMuxAddUdtaTagFunc func; } GstTagToFourcc; /* tag list tags to fourcc matching */ static const GstTagToFourcc tag_matches_mp4[] = { {FOURCC__alb, GST_TAG_ALBUM, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_soal, GST_TAG_ALBUM_SORTNAME, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__ART, GST_TAG_ARTIST, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_soar, GST_TAG_ARTIST_SORTNAME, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_aART, GST_TAG_ALBUM_ARTIST, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_soaa, GST_TAG_ALBUM_ARTIST_SORTNAME, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__swr, GST_TAG_APPLICATION_NAME, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__cmt, GST_TAG_COMMENT, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__wrt, GST_TAG_COMPOSER, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_soco, GST_TAG_COMPOSER_SORTNAME, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_tvsh, GST_TAG_SHOW_NAME, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_sosn, GST_TAG_SHOW_SORTNAME, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_tvsn, GST_TAG_SHOW_SEASON_NUMBER, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_tves, GST_TAG_SHOW_EPISODE_NUMBER, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__gen, GST_TAG_GENRE, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__nam, GST_TAG_TITLE, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_sonm, GST_TAG_TITLE_SORTNAME, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_perf, GST_TAG_PERFORMER, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__grp, GST_TAG_GROUPING, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__des, GST_TAG_DESCRIPTION, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__lyr, GST_TAG_LYRICS, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__too, GST_TAG_ENCODER, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_cprt, GST_TAG_COPYRIGHT, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_keyw, GST_TAG_KEYWORDS, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC__day, GST_TAG_DATE, NULL, gst_qt_mux_add_mp4_date}, {FOURCC_tmpo, GST_TAG_BEATS_PER_MINUTE, NULL, gst_qt_mux_add_mp4_tag}, {FOURCC_trkn, GST_TAG_TRACK_NUMBER, GST_TAG_TRACK_COUNT, gst_qt_mux_add_mp4_tag}, {FOURCC_disk, GST_TAG_ALBUM_VOLUME_NUMBER, GST_TAG_ALBUM_VOLUME_COUNT, gst_qt_mux_add_mp4_tag}, {FOURCC_covr, GST_TAG_PREVIEW_IMAGE, NULL, gst_qt_mux_add_mp4_cover}, {FOURCC_covr, GST_TAG_IMAGE, NULL, gst_qt_mux_add_mp4_cover}, {0, NULL,} }; static const GstTagToFourcc tag_matches_3gp[] = { {FOURCC_titl, GST_TAG_TITLE, NULL, gst_qt_mux_add_3gp_str}, {FOURCC_dscp, GST_TAG_DESCRIPTION, NULL, gst_qt_mux_add_3gp_str}, {FOURCC_cprt, GST_TAG_COPYRIGHT, NULL, gst_qt_mux_add_3gp_str}, {FOURCC_perf, GST_TAG_ARTIST, NULL, gst_qt_mux_add_3gp_str}, {FOURCC_auth, GST_TAG_COMPOSER, NULL, gst_qt_mux_add_3gp_str}, {FOURCC_gnre, GST_TAG_GENRE, NULL, gst_qt_mux_add_3gp_str}, {FOURCC_kywd, GST_TAG_KEYWORDS, NULL, gst_qt_mux_add_3gp_keywords}, {FOURCC_yrrc, GST_TAG_DATE, NULL, gst_qt_mux_add_3gp_date}, {FOURCC_albm, GST_TAG_ALBUM, GST_TAG_TRACK_NUMBER, gst_qt_mux_add_3gp_str}, {FOURCC_loci, GST_TAG_GEO_LOCATION_NAME, NULL, gst_qt_mux_add_3gp_location}, {FOURCC_clsf, GST_TAG_3GP_CLASSIFICATION, NULL, gst_qt_mux_add_3gp_classification}, {0, NULL,} }; /* qtdemux produces these for atoms it cannot parse */ #define GST_QT_DEMUX_PRIVATE_TAG "private-qt-tag" static void gst_qt_mux_add_xmp_tags (GstQTMux * qtmux, const GstTagList * list) { GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); GstBuffer *xmp = NULL; /* adobe specs only have 'quicktime' and 'mp4', * but I guess we can extrapolate to gpp. * Keep mj2 out for now as we don't add any tags for it yet. * If you have further info about xmp on these formats, please share */ if (qtmux_klass->format == GST_QT_MUX_FORMAT_MJ2) return; GST_DEBUG_OBJECT (qtmux, "Adding xmp tags"); if (qtmux_klass->format == GST_QT_MUX_FORMAT_QT) { xmp = gst_tag_xmp_writer_tag_list_to_xmp_buffer (GST_TAG_XMP_WRITER (qtmux), list, TRUE); if (xmp) atom_udta_add_xmp_tags (&qtmux->moov->udta, xmp); } else { AtomInfo *ainfo; /* for isom/mp4, it is a top level uuid atom */ xmp = gst_tag_xmp_writer_tag_list_to_xmp_buffer (GST_TAG_XMP_WRITER (qtmux), list, TRUE); if (xmp) { ainfo = build_uuid_xmp_atom (xmp); if (ainfo) { qtmux->extra_atoms = g_slist_prepend (qtmux->extra_atoms, ainfo); } } } if (xmp) gst_buffer_unref (xmp); } static void gst_qt_mux_add_metadata_tags (GstQTMux * qtmux, const GstTagList * list, AtomUDTA * udta) { GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); guint32 fourcc; gint i; const gchar *tag, *tag2; const GstTagToFourcc *tag_matches; switch (qtmux_klass->format) { case GST_QT_MUX_FORMAT_3GP: tag_matches = tag_matches_3gp; break; case GST_QT_MUX_FORMAT_MJ2: tag_matches = NULL; break; default: /* sort of iTunes style for mp4 and QT (?) */ tag_matches = tag_matches_mp4; break; } if (!tag_matches) return; /* Clear existing tags so we don't add them over and over */ atom_udta_clear_tags (udta); for (i = 0; tag_matches[i].fourcc; i++) { fourcc = tag_matches[i].fourcc; tag = tag_matches[i].gsttag; tag2 = tag_matches[i].gsttag2; g_assert (tag_matches[i].func); tag_matches[i].func (qtmux, list, udta, tag, tag2, fourcc); } /* add unparsed blobs if present */ if (gst_tag_exists (GST_QT_DEMUX_PRIVATE_TAG)) { guint num_tags; num_tags = gst_tag_list_get_tag_size (list, GST_QT_DEMUX_PRIVATE_TAG); for (i = 0; i < num_tags; ++i) { GstSample *sample = NULL; GstBuffer *buf; const GstStructure *s; if (!gst_tag_list_get_sample_index (list, GST_QT_DEMUX_PRIVATE_TAG, i, &sample)) continue; buf = gst_sample_get_buffer (sample); if (buf && (s = gst_sample_get_info (sample))) { const gchar *style = NULL; GstMapInfo map; gst_buffer_map (buf, &map, GST_MAP_READ); GST_DEBUG_OBJECT (qtmux, "Found private tag %d/%d; size %" G_GSIZE_FORMAT ", info %" GST_PTR_FORMAT, i, num_tags, map.size, s); if (s && (style = gst_structure_get_string (s, "style"))) { /* try to prevent some style tag ending up into another variant * (todo: make into a list if more cases) */ if ((strcmp (style, "itunes") == 0 && qtmux_klass->format == GST_QT_MUX_FORMAT_MP4) || (strcmp (style, "iso") == 0 && qtmux_klass->format == GST_QT_MUX_FORMAT_3GP)) { GST_DEBUG_OBJECT (qtmux, "Adding private tag"); atom_udta_add_blob_tag (udta, map.data, map.size); } } gst_buffer_unmap (buf, &map); } gst_sample_unref (sample); } } return; } /* * Gets the tagsetter iface taglist and puts the known tags * into the output stream */ static void gst_qt_mux_setup_metadata (GstQTMux * qtmux) { const GstTagList *tags = NULL; GSList *walk; GST_OBJECT_LOCK (qtmux); if (qtmux->tags_changed) { tags = gst_tag_setter_get_tag_list (GST_TAG_SETTER (qtmux)); qtmux->tags_changed = FALSE; } GST_OBJECT_UNLOCK (qtmux); GST_LOG_OBJECT (qtmux, "tags: %" GST_PTR_FORMAT, tags); if (tags && !gst_tag_list_is_empty (tags)) { GstTagList *copy = gst_tag_list_copy (tags); GST_DEBUG_OBJECT (qtmux, "Removing bogus tags"); gst_tag_list_remove_tag (copy, GST_TAG_VIDEO_CODEC); gst_tag_list_remove_tag (copy, GST_TAG_AUDIO_CODEC); gst_tag_list_remove_tag (copy, GST_TAG_CONTAINER_FORMAT); GST_DEBUG_OBJECT (qtmux, "Formatting tags"); gst_qt_mux_add_metadata_tags (qtmux, copy, &qtmux->moov->udta); gst_qt_mux_add_xmp_tags (qtmux, copy); gst_tag_list_unref (copy); } else { GST_DEBUG_OBJECT (qtmux, "No new tags received"); } for (walk = qtmux->sinkpads; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qpad = (GstQTPad *) cdata; GstPad *pad = qpad->collect.pad; if (qpad->tags_changed && qpad->tags) { GST_DEBUG_OBJECT (pad, "Adding tags"); gst_tag_list_remove_tag (qpad->tags, GST_TAG_CONTAINER_FORMAT); gst_qt_mux_add_metadata_tags (qtmux, qpad->tags, &qpad->trak->udta); qpad->tags_changed = FALSE; GST_DEBUG_OBJECT (pad, "Tags added"); } else { GST_DEBUG_OBJECT (pad, "No new tags received"); } } } static inline GstBuffer * _gst_buffer_new_take_data (guint8 * data, guint size) { GstBuffer *buf; buf = gst_buffer_new (); gst_buffer_append_memory (buf, gst_memory_new_wrapped (0, data, size, 0, size, data, g_free)); return buf; } static GstFlowReturn gst_qt_mux_send_buffer (GstQTMux * qtmux, GstBuffer * buf, guint64 * offset, gboolean mind_fast) { GstFlowReturn res; gsize size; g_return_val_if_fail (buf != NULL, GST_FLOW_ERROR); size = gst_buffer_get_size (buf); GST_LOG_OBJECT (qtmux, "sending buffer size %" G_GSIZE_FORMAT, size); if (mind_fast && qtmux->fast_start_file) { GstMapInfo map; gint ret; GST_LOG_OBJECT (qtmux, "to temporary file"); gst_buffer_map (buf, &map, GST_MAP_READ); ret = fwrite (map.data, sizeof (guint8), map.size, qtmux->fast_start_file); gst_buffer_unmap (buf, &map); gst_buffer_unref (buf); if (ret != size) goto write_error; else res = GST_FLOW_OK; } else { GST_LOG_OBJECT (qtmux, "downstream"); res = gst_pad_push (qtmux->srcpad, buf); } if (G_LIKELY (offset)) *offset += size; return res; /* ERRORS */ write_error: { GST_ELEMENT_ERROR (qtmux, RESOURCE, WRITE, ("Failed to write to temporary file"), GST_ERROR_SYSTEM); return GST_FLOW_ERROR; } } static gboolean gst_qt_mux_seek_to_beginning (FILE * f) { #ifdef HAVE_FSEEKO if (fseeko (f, (off_t) 0, SEEK_SET) != 0) return FALSE; #elif defined (G_OS_UNIX) || defined (G_OS_WIN32) if (lseek (fileno (f), (off_t) 0, SEEK_SET) == (off_t) - 1) return FALSE; #else if (fseek (f, (long) 0, SEEK_SET) != 0) return FALSE; #endif return TRUE; } static GstFlowReturn gst_qt_mux_send_buffered_data (GstQTMux * qtmux, guint64 * offset) { GstFlowReturn ret = GST_FLOW_OK; GstBuffer *buf = NULL; if (fflush (qtmux->fast_start_file)) goto flush_failed; if (!gst_qt_mux_seek_to_beginning (qtmux->fast_start_file)) goto seek_failed; /* hm, this could all take a really really long time, * but there may not be another way to get moov atom first * (somehow optimize copy?) */ GST_DEBUG_OBJECT (qtmux, "Sending buffered data"); while (ret == GST_FLOW_OK) { const int bufsize = 4096; GstMapInfo map; gsize size; buf = gst_buffer_new_and_alloc (bufsize); gst_buffer_map (buf, &map, GST_MAP_WRITE); size = fread (map.data, sizeof (guint8), bufsize, qtmux->fast_start_file); if (size == 0) { gst_buffer_unmap (buf, &map); break; } GST_LOG_OBJECT (qtmux, "Pushing buffered buffer of size %d", (gint) size); gst_buffer_unmap (buf, &map); if (size != bufsize) gst_buffer_set_size (buf, size); ret = gst_qt_mux_send_buffer (qtmux, buf, offset, FALSE); buf = NULL; } if (buf) gst_buffer_unref (buf); if (ftruncate (fileno (qtmux->fast_start_file), 0)) goto seek_failed; if (!gst_qt_mux_seek_to_beginning (qtmux->fast_start_file)) goto seek_failed; return ret; /* ERRORS */ flush_failed: { GST_ELEMENT_ERROR (qtmux, RESOURCE, WRITE, ("Failed to flush temporary file"), GST_ERROR_SYSTEM); ret = GST_FLOW_ERROR; goto fail; } seek_failed: { GST_ELEMENT_ERROR (qtmux, RESOURCE, SEEK, ("Failed to seek temporary file"), GST_ERROR_SYSTEM); ret = GST_FLOW_ERROR; goto fail; } fail: { /* clear descriptor so we don't remove temp file later on, * might be possible to recover */ fclose (qtmux->fast_start_file); qtmux->fast_start_file = NULL; return ret; } } /* * Sends the initial mdat atom fields (size fields and fourcc type), * the subsequent buffers are considered part of it's data. * As we can't predict the amount of data that we are going to place in mdat * we need to record the position of the size field in the stream so we can * seek back to it later and update when the streams have finished. */ static GstFlowReturn gst_qt_mux_send_mdat_header (GstQTMux * qtmux, guint64 * off, guint64 size, gboolean extended, gboolean fsync_after) { GstBuffer *buf; GstMapInfo map; GST_DEBUG_OBJECT (qtmux, "Sending mdat's atom header, " "size %" G_GUINT64_FORMAT, size); /* if the qtmux state is EOS, really write the mdat, otherwise * allow size == 0 for a placeholder atom */ if (qtmux->state == GST_QT_MUX_STATE_EOS || size > 0) size += 8; if (extended) { gboolean large_file = (size > MDAT_LARGE_FILE_LIMIT); /* Always write 16-bytes, but put a free atom first * if the size is < 4GB. */ buf = gst_buffer_new_and_alloc (16); gst_buffer_map (buf, &map, GST_MAP_WRITE); if (large_file) { /* Write extended mdat header and large_size field */ GST_WRITE_UINT32_BE (map.data, 1); GST_WRITE_UINT32_LE (map.data + 4, FOURCC_mdat); GST_WRITE_UINT64_BE (map.data + 8, size + 8); } else { /* Write an empty free atom, then standard 32-bit mdat */ GST_WRITE_UINT32_BE (map.data, 8); GST_WRITE_UINT32_LE (map.data + 4, FOURCC_free); GST_WRITE_UINT32_BE (map.data + 8, size); GST_WRITE_UINT32_LE (map.data + 12, FOURCC_mdat); } gst_buffer_unmap (buf, &map); } else { buf = gst_buffer_new_and_alloc (8); gst_buffer_map (buf, &map, GST_MAP_WRITE); /* Vanilla 32-bit mdat */ GST_WRITE_UINT32_BE (map.data, size); GST_WRITE_UINT32_LE (map.data + 4, FOURCC_mdat); gst_buffer_unmap (buf, &map); } GST_LOG_OBJECT (qtmux, "Pushing mdat header"); if (fsync_after) GST_BUFFER_FLAG_SET (buf, GST_BUFFER_FLAG_SYNC_AFTER); return gst_qt_mux_send_buffer (qtmux, buf, off, FALSE); } /* * We get the position of the mdat size field, seek back to it * and overwrite with the real value */ static GstFlowReturn gst_qt_mux_update_mdat_size (GstQTMux * qtmux, guint64 mdat_pos, guint64 mdat_size, guint64 * offset, gboolean fsync_after) { GstSegment segment; /* We must have recorded the mdat position for this to work */ g_assert (mdat_pos != 0); /* seek and rewrite the header */ gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = mdat_pos; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); return gst_qt_mux_send_mdat_header (qtmux, offset, mdat_size, TRUE, fsync_after); } static GstFlowReturn gst_qt_mux_send_ftyp (GstQTMux * qtmux, guint64 * off) { GstBuffer *buf; guint64 size = 0, offset = 0; guint8 *data = NULL; GST_DEBUG_OBJECT (qtmux, "Sending ftyp atom"); if (!atom_ftyp_copy_data (qtmux->ftyp, &data, &size, &offset)) goto serialize_error; buf = _gst_buffer_new_take_data (data, offset); GST_LOG_OBJECT (qtmux, "Pushing ftyp"); return gst_qt_mux_send_buffer (qtmux, buf, off, FALSE); /* ERRORS */ serialize_error: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Failed to serialize ftyp")); return GST_FLOW_ERROR; } } static void gst_qt_mux_prepare_ftyp (GstQTMux * qtmux, AtomFTYP ** p_ftyp, GstBuffer ** p_prefix) { GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); guint32 major, version; GList *comp; GstBuffer *prefix = NULL; AtomFTYP *ftyp = NULL; GST_DEBUG_OBJECT (qtmux, "Preparing ftyp and possible prefix atom"); /* init and send context and ftyp based on current property state */ gst_qt_mux_map_format_to_header (qtmux_klass->format, &prefix, &major, &version, &comp, qtmux->moov, qtmux->longest_chunk, qtmux->fast_start_file != NULL); ftyp = atom_ftyp_new (qtmux->context, major, version, comp); if (comp) g_list_free (comp); if (prefix) { if (p_prefix) *p_prefix = prefix; else gst_buffer_unref (prefix); } *p_ftyp = ftyp; } static GstFlowReturn gst_qt_mux_prepare_and_send_ftyp (GstQTMux * qtmux) { GstFlowReturn ret = GST_FLOW_OK; GstBuffer *prefix = NULL; GST_DEBUG_OBJECT (qtmux, "Preparing to send ftyp atom"); /* init and send context and ftyp based on current property state */ if (qtmux->ftyp) { atom_ftyp_free (qtmux->ftyp); qtmux->ftyp = NULL; } gst_qt_mux_prepare_ftyp (qtmux, &qtmux->ftyp, &prefix); if (prefix) { ret = gst_qt_mux_send_buffer (qtmux, prefix, &qtmux->header_size, FALSE); if (ret != GST_FLOW_OK) return ret; } return gst_qt_mux_send_ftyp (qtmux, &qtmux->header_size); } static void gst_qt_mux_set_header_on_caps (GstQTMux * mux, GstBuffer * buf) { GstStructure *structure; GValue array = { 0 }; GValue value = { 0 }; GstCaps *caps, *tcaps; tcaps = gst_pad_get_current_caps (mux->srcpad); caps = gst_caps_copy (tcaps); gst_caps_unref (tcaps); structure = gst_caps_get_structure (caps, 0); g_value_init (&array, GST_TYPE_ARRAY); GST_BUFFER_FLAG_SET (buf, GST_BUFFER_FLAG_HEADER); g_value_init (&value, GST_TYPE_BUFFER); gst_value_take_buffer (&value, gst_buffer_ref (buf)); gst_value_array_append_value (&array, &value); g_value_unset (&value); gst_structure_set_value (structure, "streamheader", &array); g_value_unset (&array); gst_pad_set_caps (mux->srcpad, caps); gst_caps_unref (caps); } /* * Write out a free space atom. The offset is adjusted by the full * size, but a smaller buffer is sent */ static GstFlowReturn gst_qt_mux_send_free_atom (GstQTMux * qtmux, guint64 * off, guint32 size, gboolean fsync_after) { Atom *node_header; GstBuffer *buf; guint8 *data = NULL; guint64 offset = 0, bsize = 0; GstFlowReturn ret; GST_DEBUG_OBJECT (qtmux, "Sending free atom header of size %u", size); /* We can't make a free space atom smaller than the header */ if (size < 8) goto too_small; node_header = g_malloc0 (sizeof (Atom)); node_header->type = FOURCC_free; node_header->size = size; bsize = offset = 0; if (atom_copy_data (node_header, &data, &bsize, &offset) == 0) goto serialize_error; buf = _gst_buffer_new_take_data (data, offset); g_free (node_header); if (fsync_after) GST_BUFFER_FLAG_SET (buf, GST_BUFFER_FLAG_SYNC_AFTER); GST_LOG_OBJECT (qtmux, "Pushing free atom"); ret = gst_qt_mux_send_buffer (qtmux, buf, off, FALSE); if (off) { GstSegment segment; *off += size - 8; /* Make sure downstream position ends up at the end of this free box */ gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = *off; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); } return ret; /* ERRORS */ too_small: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Not enough free reserved space")); return GST_FLOW_ERROR; } serialize_error: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Failed to serialize mdat")); g_free (node_header); return GST_FLOW_ERROR; } } static void gst_qt_mux_configure_moov (GstQTMux * qtmux) { gboolean fragmented = FALSE; guint32 timescale; GST_OBJECT_LOCK (qtmux); timescale = qtmux->timescale; if (qtmux->mux_mode == GST_QT_MUX_MODE_FRAGMENTED || qtmux->mux_mode == GST_QT_MUX_MODE_FRAGMENTED_STREAMABLE) fragmented = TRUE; GST_OBJECT_UNLOCK (qtmux); /* inform lower layers of our property wishes, and determine duration. * Let moov take care of this using its list of traks; * so that released pads are also included */ GST_DEBUG_OBJECT (qtmux, "Updating timescale to %" G_GUINT32_FORMAT, timescale); atom_moov_update_timescale (qtmux->moov, timescale); atom_moov_set_fragmented (qtmux->moov, fragmented); atom_moov_update_duration (qtmux->moov); } static GstFlowReturn gst_qt_mux_send_moov (GstQTMux * qtmux, guint64 * _offset, guint64 padded_moov_size, gboolean mind_fast, gboolean fsync_after) { guint64 offset = 0, size = 0; guint8 *data; GstBuffer *buf; GstFlowReturn ret = GST_FLOW_OK; GSList *walk; guint64 current_time = atoms_get_current_qt_time (); /* update modification times */ qtmux->moov->mvhd.time_info.modification_time = current_time; for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; qtpad->trak->mdia.mdhd.time_info.modification_time = current_time; qtpad->trak->tkhd.modification_time = current_time; } /* serialize moov */ offset = size = 0; data = NULL; GST_LOG_OBJECT (qtmux, "Copying movie header into buffer"); if (!atom_moov_copy_data (qtmux->moov, &data, &size, &offset)) goto serialize_error; qtmux->last_moov_size = offset; /* Check we have enough reserved space for this and a Free atom */ if (padded_moov_size > 0 && offset + 8 > padded_moov_size) goto too_small_reserved; buf = _gst_buffer_new_take_data (data, offset); GST_DEBUG_OBJECT (qtmux, "Pushing moov atoms"); /* If at EOS, this is the final moov, put in the streamheader * (apparently used by a flumotion util) */ if (qtmux->state == GST_QT_MUX_STATE_EOS) gst_qt_mux_set_header_on_caps (qtmux, buf); if (fsync_after) GST_BUFFER_FLAG_SET (buf, GST_BUFFER_FLAG_SYNC_AFTER); ret = gst_qt_mux_send_buffer (qtmux, buf, _offset, mind_fast); /* Write out a free atom if needed */ if (ret == GST_FLOW_OK && offset < padded_moov_size) { GST_LOG_OBJECT (qtmux, "Writing out free atom of size %u", (guint32) (padded_moov_size - offset)); ret = gst_qt_mux_send_free_atom (qtmux, _offset, padded_moov_size - offset, fsync_after); } return ret; too_small_reserved: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, ("Not enough free reserved header space"), ("Needed %" G_GUINT64_FORMAT " bytes, reserved %" G_GUINT64_FORMAT, offset, padded_moov_size)); return GST_FLOW_ERROR; } serialize_error: { g_free (data); return GST_FLOW_ERROR; } } /* either calculates size of extra atoms or pushes them */ static GstFlowReturn gst_qt_mux_send_extra_atoms (GstQTMux * qtmux, gboolean send, guint64 * offset, gboolean mind_fast) { GSList *walk; guint64 loffset = 0, size = 0; guint8 *data; GstFlowReturn ret = GST_FLOW_OK; for (walk = qtmux->extra_atoms; walk; walk = g_slist_next (walk)) { AtomInfo *ainfo = (AtomInfo *) walk->data; loffset = size = 0; data = NULL; if (!ainfo->copy_data_func (ainfo->atom, send ? &data : NULL, &size, &loffset)) goto serialize_error; if (send) { GstBuffer *buf; GST_DEBUG_OBJECT (qtmux, "Pushing extra top-level atom %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (ainfo->atom->type)); buf = _gst_buffer_new_take_data (data, loffset); ret = gst_qt_mux_send_buffer (qtmux, buf, offset, FALSE); if (ret != GST_FLOW_OK) break; } else { if (offset) *offset += loffset; } } return ret; serialize_error: { g_free (data); return GST_FLOW_ERROR; } } static gboolean gst_qt_mux_downstream_is_seekable (GstQTMux * qtmux) { gboolean seekable = FALSE; GstQuery *query = gst_query_new_seeking (GST_FORMAT_BYTES); if (gst_pad_peer_query (qtmux->srcpad, query)) { gst_query_parse_seeking (query, NULL, &seekable, NULL, NULL); GST_INFO_OBJECT (qtmux, "downstream is %sseekable", seekable ? "" : "not "); } else { /* have to assume seeking is not supported if query not handled downstream */ GST_WARNING_OBJECT (qtmux, "downstream did not handle seeking query"); seekable = FALSE; } gst_query_unref (query); return seekable; } static void gst_qt_mux_prepare_moov_recovery (GstQTMux * qtmux) { GSList *walk; gboolean fail = FALSE; AtomFTYP *ftyp = NULL; GstBuffer *prefix = NULL; GST_DEBUG_OBJECT (qtmux, "Opening moov recovery file: %s", qtmux->moov_recov_file_path); qtmux->moov_recov_file = g_fopen (qtmux->moov_recov_file_path, "wb+"); if (qtmux->moov_recov_file == NULL) { GST_WARNING_OBJECT (qtmux, "Failed to open moov recovery file in %s", qtmux->moov_recov_file_path); return; } gst_qt_mux_prepare_ftyp (qtmux, &ftyp, &prefix); if (!atoms_recov_write_headers (qtmux->moov_recov_file, ftyp, prefix, qtmux->moov, qtmux->timescale, g_slist_length (qtmux->sinkpads))) { GST_WARNING_OBJECT (qtmux, "Failed to write moov recovery file " "headers"); goto fail; } atom_ftyp_free (ftyp); if (prefix) gst_buffer_unref (prefix); for (walk = qtmux->sinkpads; walk && !fail; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qpad = (GstQTPad *) cdata; /* write info for each stream */ fail = atoms_recov_write_trak_info (qtmux->moov_recov_file, qpad->trak); if (fail) { GST_WARNING_OBJECT (qtmux, "Failed to write trak info to recovery " "file"); break; } } return; fail: /* cleanup */ fclose (qtmux->moov_recov_file); qtmux->moov_recov_file = NULL; } static guint64 prefill_get_block_index (GstQTMux * qtmux, GstQTPad * qpad) { switch (qpad->fourcc) { case FOURCC_apch: case FOURCC_apcn: case FOURCC_apcs: case FOURCC_apco: case FOURCC_ap4h: case FOURCC_ap4x: case FOURCC_c608: case FOURCC_c708: return qpad->sample_offset; case FOURCC_sowt: case FOURCC_twos: return gst_util_uint64_scale_ceil (qpad->sample_offset, qpad->expected_sample_duration_n, qpad->expected_sample_duration_d * atom_trak_get_timescale (qpad->trak)); default: return -1; } } static guint prefill_get_sample_size (GstQTMux * qtmux, GstQTPad * qpad) { switch (qpad->fourcc) { case FOURCC_apch: if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 480) { return 300000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 576) { return 350000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 720) { return 525000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 1080) { return 1050000; } else { return 4150000; } break; case FOURCC_apcn: if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 480) { return 200000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 576) { return 250000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 720) { return 350000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 1080) { return 700000; } else { return 2800000; } break; case FOURCC_apcs: if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 480) { return 150000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 576) { return 200000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 720) { return 250000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 1080) { return 500000; } else { return 2800000; } break; case FOURCC_apco: if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 480) { return 80000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 576) { return 100000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 720) { return 150000; } else if (((SampleTableEntryMP4V *) qpad->trak_ste)->height <= 1080) { return 250000; } else { return 900000; } break; case FOURCC_c608: /* We always write both cdat and cdt2 atom in prefill mode */ return 20; case FOURCC_c708: /* We're cheating a bit by always allocating 256 bytes plus 8 bytes for the atom header * even if we use less */ return 256 + 8; case FOURCC_sowt: case FOURCC_twos:{ guint64 block_idx; guint64 next_sample_offset; block_idx = prefill_get_block_index (qtmux, qpad); next_sample_offset = gst_util_uint64_scale (block_idx + 1, qpad->expected_sample_duration_d * atom_trak_get_timescale (qpad->trak), qpad->expected_sample_duration_n); return (next_sample_offset - qpad->sample_offset) * qpad->sample_size; } case FOURCC_ap4h: case FOURCC_ap4x: default: GST_ERROR_OBJECT (qtmux, "unsupported codec for pre-filling"); return -1; } return -1; } static GstClockTime prefill_get_next_timestamp (GstQTMux * qtmux, GstQTPad * qpad) { switch (qpad->fourcc) { case FOURCC_apch: case FOURCC_apcn: case FOURCC_apcs: case FOURCC_apco: case FOURCC_ap4h: case FOURCC_ap4x: case FOURCC_c608: case FOURCC_c708: return gst_util_uint64_scale (qpad->sample_offset + 1, qpad->expected_sample_duration_d * GST_SECOND, qpad->expected_sample_duration_n); case FOURCC_sowt: case FOURCC_twos:{ guint64 block_idx; guint64 next_sample_offset; block_idx = prefill_get_block_index (qtmux, qpad); next_sample_offset = gst_util_uint64_scale (block_idx + 1, qpad->expected_sample_duration_d * atom_trak_get_timescale (qpad->trak), qpad->expected_sample_duration_n); return gst_util_uint64_scale (next_sample_offset, GST_SECOND, atom_trak_get_timescale (qpad->trak)); } default: GST_ERROR_OBJECT (qtmux, "unsupported codec for pre-filling"); return -1; } return -1; } static GstBuffer * prefill_raw_audio_prepare_buf_func (GstQTPad * qtpad, GstBuffer * buf, GstQTMux * qtmux) { guint64 block_idx; guint64 nsamples; GstClockTime input_timestamp; guint64 input_timestamp_distance; if (buf) gst_adapter_push (qtpad->raw_audio_adapter, buf); block_idx = gst_util_uint64_scale_ceil (qtpad->raw_audio_adapter_offset, qtpad->expected_sample_duration_n, qtpad->expected_sample_duration_d * atom_trak_get_timescale (qtpad->trak)); nsamples = gst_util_uint64_scale (block_idx + 1, qtpad->expected_sample_duration_d * atom_trak_get_timescale (qtpad->trak), qtpad->expected_sample_duration_n) - qtpad->raw_audio_adapter_offset; if ((!GST_COLLECT_PADS_STATE_IS_SET (&qtpad->collect, GST_COLLECT_PADS_STATE_EOS) && gst_adapter_available (qtpad->raw_audio_adapter) < nsamples * qtpad->sample_size) || gst_adapter_available (qtpad->raw_audio_adapter) == 0) { return NULL; } input_timestamp = gst_adapter_prev_pts (qtpad->raw_audio_adapter, &input_timestamp_distance); if (input_timestamp != GST_CLOCK_TIME_NONE) input_timestamp += gst_util_uint64_scale (input_timestamp_distance, GST_SECOND, qtpad->sample_size * atom_trak_get_timescale (qtpad->trak)); buf = gst_adapter_take_buffer (qtpad->raw_audio_adapter, !GST_COLLECT_PADS_STATE_IS_SET (&qtpad->collect, GST_COLLECT_PADS_STATE_EOS) ? nsamples * qtpad->sample_size : gst_adapter_available (qtpad->raw_audio_adapter)); GST_BUFFER_PTS (buf) = input_timestamp; GST_BUFFER_DTS (buf) = GST_CLOCK_TIME_NONE; GST_BUFFER_DURATION (buf) = GST_CLOCK_TIME_NONE; qtpad->raw_audio_adapter_offset += nsamples; /* Check if we have yet another block of raw audio in the adapter */ nsamples = gst_util_uint64_scale (block_idx + 2, qtpad->expected_sample_duration_d * atom_trak_get_timescale (qtpad->trak), qtpad->expected_sample_duration_n) - qtpad->raw_audio_adapter_offset; if (gst_adapter_available (qtpad->raw_audio_adapter) >= nsamples * qtpad->sample_size) { input_timestamp = gst_adapter_prev_pts (qtpad->raw_audio_adapter, &input_timestamp_distance); if (input_timestamp != GST_CLOCK_TIME_NONE) input_timestamp += gst_util_uint64_scale (input_timestamp_distance, GST_SECOND, qtpad->sample_size * atom_trak_get_timescale (qtpad->trak)); qtpad->raw_audio_adapter_pts = input_timestamp; } else { qtpad->raw_audio_adapter_pts = GST_CLOCK_TIME_NONE; } return buf; } static void find_video_sample_duration (GstQTMux * qtmux, guint * dur_n, guint * dur_d) { GSList *walk; /* Find the (first) video track and assume that we have to output * in that size */ for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *tmp_qpad = (GstQTPad *) cdata; if (tmp_qpad->trak->is_video) { *dur_n = tmp_qpad->expected_sample_duration_n; *dur_d = tmp_qpad->expected_sample_duration_d; break; } } if (walk == NULL) { GST_INFO_OBJECT (qtmux, "Found no video framerate, using 40ms audio buffers"); *dur_n = 25; *dur_d = 1; } } /* Called when all pads are prerolled to adjust and */ static gboolean prefill_update_sample_size (GstQTMux * qtmux, GstQTPad * qpad) { switch (qpad->fourcc) { case FOURCC_apch: case FOURCC_apcn: case FOURCC_apcs: case FOURCC_apco: case FOURCC_ap4h: case FOURCC_ap4x: { guint sample_size = prefill_get_sample_size (qtmux, qpad); atom_trak_set_constant_size_samples (qpad->trak, sample_size); return TRUE; } case FOURCC_c608: case FOURCC_c708: { guint sample_size = prefill_get_sample_size (qtmux, qpad); /* We need a "valid" duration */ find_video_sample_duration (qtmux, &qpad->expected_sample_duration_n, &qpad->expected_sample_duration_d); atom_trak_set_constant_size_samples (qpad->trak, sample_size); return TRUE; } case FOURCC_sowt: case FOURCC_twos:{ find_video_sample_duration (qtmux, &qpad->expected_sample_duration_n, &qpad->expected_sample_duration_d); /* Set a prepare_buf_func that ensures this */ qpad->prepare_buf_func = prefill_raw_audio_prepare_buf_func; qpad->raw_audio_adapter = gst_adapter_new (); qpad->raw_audio_adapter_offset = 0; qpad->raw_audio_adapter_pts = GST_CLOCK_TIME_NONE; return TRUE; } default: return TRUE; } } /* Only called at startup when doing the "fake" iteration of all tracks in order * to prefill the sample tables in the header. */ static GstQTPad * find_best_pad_prefill_start (GstQTMux * qtmux) { GSList *walk; GstQTPad *best_pad = NULL; /* If interleave limits have been specified and the current pad is within * those interleave limits, pick that one, otherwise let's try to figure out * the next best one. */ if (qtmux->current_pad && (qtmux->interleave_bytes != 0 || qtmux->interleave_time != 0) && (qtmux->interleave_bytes == 0 || qtmux->current_chunk_size <= qtmux->interleave_bytes) && (qtmux->interleave_time == 0 || qtmux->current_chunk_duration <= qtmux->interleave_time) && qtmux->mux_mode != GST_QT_MUX_MODE_FRAGMENTED && qtmux->mux_mode != GST_QT_MUX_MODE_FRAGMENTED_STREAMABLE) { if (qtmux->current_pad->total_duration < qtmux->reserved_max_duration) { best_pad = qtmux->current_pad; } } else if (qtmux->collect->data->next) { /* Attempt to try another pad if we have one. Otherwise use the only pad * present */ best_pad = qtmux->current_pad = NULL; } /* The next best pad is the one which has the lowest timestamp and hasn't * exceeded the reserved max duration */ if (!best_pad) { GstClockTime best_time = GST_CLOCK_TIME_NONE; for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; GstClockTime timestamp; if (qtpad->total_duration >= qtmux->reserved_max_duration) continue; timestamp = qtpad->total_duration; if (best_pad == NULL || !GST_CLOCK_TIME_IS_VALID (best_time) || timestamp < best_time) { best_pad = qtpad; best_time = timestamp; } } } return best_pad; } /* Called when starting the file in prefill_mode to figure out all the entries * of the header based on the input stream and reserved maximum duration. * * The _actual_ header (i.e. with the proper duration and trimmed sample tables) * will be updated and written on EOS. */ static gboolean gst_qt_mux_prefill_samples (GstQTMux * qtmux) { GstQTPad *qpad; GSList *walk; GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); /* Update expected sample sizes/durations as needed, this is for raw * audio where samples are actual audio samples. */ for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qpad = (GstQTPad *) cdata; if (!prefill_update_sample_size (qtmux, qpad)) return FALSE; } if (qtmux_klass->format == GST_QT_MUX_FORMAT_QT) { /* For the first sample check/update timecode as needed. We do that before * all actual samples as the code in gst_qt_mux_add_buffer() does it with * initial buffer directly, not with last_buf */ for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qpad = (GstQTPad *) cdata; GstBuffer *buffer = gst_collect_pads_peek (qtmux->collect, (GstCollectData *) qpad); GstVideoTimeCodeMeta *tc_meta; if (buffer && (tc_meta = gst_buffer_get_video_time_code_meta (buffer)) && qpad->trak->is_video) { GstVideoTimeCode *tc = &tc_meta->tc; qpad->tc_trak = atom_trak_new (qtmux->context); atom_moov_add_trak (qtmux->moov, qpad->tc_trak); qpad->trak->tref = atom_tref_new (FOURCC_tmcd); atom_tref_add_entry (qpad->trak->tref, qpad->tc_trak->tkhd.track_ID); atom_trak_set_timecode_type (qpad->tc_trak, qtmux->context, qpad->trak->mdia.mdhd.time_info.timescale, tc); atom_trak_add_samples (qpad->tc_trak, 1, 1, 4, qtmux->mdat_size, FALSE, 0); qpad->tc_pos = qtmux->mdat_size; qpad->first_tc = gst_video_time_code_copy (tc); qpad->first_pts = GST_BUFFER_PTS (buffer); qtmux->current_chunk_offset = -1; qtmux->current_chunk_size = 0; qtmux->current_chunk_duration = 0; qtmux->mdat_size += 4; } if (buffer) gst_buffer_unref (buffer); } } while ((qpad = find_best_pad_prefill_start (qtmux))) { GstClockTime timestamp, next_timestamp, duration; guint nsamples, sample_size; guint64 chunk_offset; gint64 scaled_duration; gint64 pts_offset = 0; gboolean sync = FALSE; TrakBufferEntryInfo sample_entry; sample_size = prefill_get_sample_size (qtmux, qpad); if (sample_size == -1) { return FALSE; } if (!qpad->samples) qpad->samples = g_array_new (FALSE, FALSE, sizeof (TrakBufferEntryInfo)); timestamp = qpad->total_duration; next_timestamp = prefill_get_next_timestamp (qtmux, qpad); duration = next_timestamp - timestamp; if (qpad->first_ts == GST_CLOCK_TIME_NONE) qpad->first_ts = timestamp; if (qpad->first_dts == GST_CLOCK_TIME_NONE) qpad->first_dts = timestamp; if (qtmux->current_pad != qpad || qtmux->current_chunk_offset == -1) { qtmux->current_pad = qpad; if (qtmux->current_chunk_offset == -1) qtmux->current_chunk_offset = qtmux->mdat_size; else qtmux->current_chunk_offset += qtmux->current_chunk_size; qtmux->current_chunk_size = 0; qtmux->current_chunk_duration = 0; } if (qpad->sample_size) nsamples = sample_size / qpad->sample_size; else nsamples = 1; qpad->last_dts = timestamp; scaled_duration = gst_util_uint64_scale_round (timestamp + duration, atom_trak_get_timescale (qpad->trak), GST_SECOND) - gst_util_uint64_scale_round (timestamp, atom_trak_get_timescale (qpad->trak), GST_SECOND); qtmux->current_chunk_size += sample_size; qtmux->current_chunk_duration += duration; qpad->total_bytes += sample_size; chunk_offset = qtmux->current_chunk_offset; /* I-frame only, no frame reordering */ sync = FALSE; pts_offset = 0; if (qtmux->current_chunk_duration > qtmux->longest_chunk || !GST_CLOCK_TIME_IS_VALID (qtmux->longest_chunk)) { qtmux->longest_chunk = qtmux->current_chunk_duration; } sample_entry.track_id = qpad->trak->tkhd.track_ID; sample_entry.nsamples = nsamples; sample_entry.delta = scaled_duration / nsamples; sample_entry.size = sample_size / nsamples; sample_entry.chunk_offset = chunk_offset; sample_entry.pts_offset = pts_offset; sample_entry.sync = sync; sample_entry.do_pts = TRUE; g_array_append_val (qpad->samples, sample_entry); atom_trak_add_samples (qpad->trak, nsamples, scaled_duration / nsamples, sample_size / nsamples, chunk_offset, sync, pts_offset); qpad->total_duration = next_timestamp; qtmux->mdat_size += sample_size; qpad->sample_offset += nsamples; } return TRUE; } static GstFlowReturn gst_qt_mux_start_file (GstQTMux * qtmux) { GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); GstFlowReturn ret = GST_FLOW_OK; GstCaps *caps; GstSegment segment; gchar s_id[32]; GstClockTime reserved_max_duration; guint reserved_bytes_per_sec_per_trak; GSList *walk; GST_DEBUG_OBJECT (qtmux, "starting file"); GST_OBJECT_LOCK (qtmux); reserved_max_duration = qtmux->reserved_max_duration; reserved_bytes_per_sec_per_trak = qtmux->reserved_bytes_per_sec_per_trak; GST_OBJECT_UNLOCK (qtmux); /* stream-start (FIXME: create id based on input ids) */ g_snprintf (s_id, sizeof (s_id), "qtmux-%08x", g_random_int ()); gst_pad_push_event (qtmux->srcpad, gst_event_new_stream_start (s_id)); caps = gst_caps_copy (gst_pad_get_pad_template_caps (qtmux->srcpad)); /* qtmux has structure with and without variant, remove all but the first */ while (gst_caps_get_size (caps) > 1) gst_caps_remove_structure (caps, 1); gst_pad_set_caps (qtmux->srcpad, caps); gst_caps_unref (caps); /* Default is 'normal' mode */ qtmux->mux_mode = GST_QT_MUX_MODE_MOOV_AT_END; /* Require a sensible fragment duration when muxing * using the ISML muxer */ if (qtmux_klass->format == GST_QT_MUX_FORMAT_ISML && qtmux->fragment_duration == 0) goto invalid_isml; if (qtmux->fragment_duration > 0) { if (qtmux->streamable) qtmux->mux_mode = GST_QT_MUX_MODE_FRAGMENTED_STREAMABLE; else qtmux->mux_mode = GST_QT_MUX_MODE_FRAGMENTED; } else if (qtmux->fast_start) { qtmux->mux_mode = GST_QT_MUX_MODE_FAST_START; } else if (reserved_max_duration != GST_CLOCK_TIME_NONE) { if (qtmux->reserved_prefill) qtmux->mux_mode = GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL; else qtmux->mux_mode = GST_QT_MUX_MODE_ROBUST_RECORDING; } switch (qtmux->mux_mode) { case GST_QT_MUX_MODE_MOOV_AT_END: case GST_QT_MUX_MODE_ROBUST_RECORDING: /* We have to be able to seek to rewrite the mdat header, or any * moov atom we write will not be visible in the file, because an * MDAT with 0 as the size covers the rest of the file. A file * with no moov is not playable, so error out now. */ if (!gst_qt_mux_downstream_is_seekable (qtmux)) { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, ("Downstream is not seekable - will not be able to create a playable file"), (NULL)); return GST_FLOW_ERROR; } if (qtmux->reserved_moov_update_period == GST_CLOCK_TIME_NONE) { GST_WARNING_OBJECT (qtmux, "Robust muxing requires reserved-moov-update-period to be set"); } break; case GST_QT_MUX_MODE_FAST_START: case GST_QT_MUX_MODE_FRAGMENTED_STREAMABLE: break; /* Don't need seekability, ignore */ case GST_QT_MUX_MODE_FRAGMENTED: if (!gst_qt_mux_downstream_is_seekable (qtmux)) { GST_WARNING_OBJECT (qtmux, "downstream is not seekable, but " "streamable=false. Will ignore that and create streamable output " "instead"); qtmux->streamable = TRUE; g_object_notify (G_OBJECT (qtmux), "streamable"); } break; case GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL: if (!gst_qt_mux_downstream_is_seekable (qtmux)) { GST_WARNING_OBJECT (qtmux, "downstream is not seekable, will not be able " "to trim samples table at the end if less than reserved-duration is " "recorded"); } break; } /* let downstream know we think in BYTES and expect to do seeking later on */ gst_segment_init (&segment, GST_FORMAT_BYTES); gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); GST_OBJECT_LOCK (qtmux); if (qtmux->timescale == 0) { guint32 suggested_timescale = 0; GSList *walk; /* Calculate a reasonable timescale for the moov: * If there is video, it is the biggest video track timescale or an even * multiple of it if it's smaller than 1800. * Otherwise it is 1800 */ for (walk = qtmux->sinkpads; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qpad = (GstQTPad *) cdata; if (!qpad->trak) continue; /* not video */ if (!qpad->trak->mdia.minf.vmhd) continue; suggested_timescale = MAX (qpad->trak->mdia.mdhd.time_info.timescale, suggested_timescale); } if (suggested_timescale == 0) suggested_timescale = 1800; while (suggested_timescale < 1800) suggested_timescale *= 2; qtmux->timescale = suggested_timescale; } /* Set width/height/timescale of any closed caption tracks to that of the * first video track */ { guint video_width = 0, video_height = 0; guint32 video_timescale = 0; GSList *walk; for (walk = qtmux->sinkpads; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qpad = (GstQTPad *) cdata; if (!qpad->trak) continue; /* Not closed caption */ if (qpad->trak->mdia.hdlr.handler_type != FOURCC_clcp) continue; if (video_width == 0 || video_height == 0 || video_timescale == 0) { GSList *walk2; for (walk2 = qtmux->sinkpads; walk2; walk2 = g_slist_next (walk2)) { GstCollectData *cdata2 = (GstCollectData *) walk2->data; GstQTPad *qpad2 = (GstQTPad *) cdata2; if (!qpad2->trak) continue; /* not video */ if (!qpad2->trak->mdia.minf.vmhd) continue; video_width = qpad2->trak->tkhd.width; video_height = qpad2->trak->tkhd.height; video_timescale = qpad2->trak->mdia.mdhd.time_info.timescale; } } qpad->trak->tkhd.width = video_width << 16; qpad->trak->tkhd.height = video_height << 16; qpad->trak->mdia.mdhd.time_info.timescale = video_timescale; } } /* initialize our moov recovery file */ if (qtmux->moov_recov_file_path) { gst_qt_mux_prepare_moov_recovery (qtmux); } /* Make sure the first time we update the moov, we'll * include any tagsetter tags */ qtmux->tags_changed = TRUE; GST_OBJECT_UNLOCK (qtmux); /* * send mdat header if already needed, and mark position for later update. * We don't send ftyp now if we are on fast start mode, because we can * better fine tune using the information we gather to create the whole moov * atom. */ switch (qtmux->mux_mode) { case GST_QT_MUX_MODE_MOOV_AT_END: ret = gst_qt_mux_prepare_and_send_ftyp (qtmux); if (ret != GST_FLOW_OK) break; /* Store this as the mdat offset for later updating * when we write the moov */ qtmux->mdat_pos = qtmux->header_size; /* extended atom in case we go over 4GB while writing and need * the full 64-bit atom */ ret = gst_qt_mux_send_mdat_header (qtmux, &qtmux->header_size, 0, TRUE, FALSE); break; case GST_QT_MUX_MODE_ROBUST_RECORDING: ret = gst_qt_mux_prepare_and_send_ftyp (qtmux); if (ret != GST_FLOW_OK) break; /* Pad ftyp out to an 8-byte boundary before starting the moov * ping pong region. It should be well less than 1 disk sector, * unless there's a bajillion compatible types listed, * but let's be sure the free atom doesn't cross a sector * boundary anyway */ if (qtmux->header_size % 8) { /* Extra 8 bytes for the padding free atom header */ guint padding = (guint) (16 - (qtmux->header_size % 8)); GST_LOG_OBJECT (qtmux, "Rounding ftyp by %u bytes", padding); ret = gst_qt_mux_send_free_atom (qtmux, &qtmux->header_size, padding, FALSE); if (ret != GST_FLOW_OK) return ret; } /* Store this as the moov offset for later updating. * We record mdat position below */ qtmux->moov_pos = qtmux->header_size; /* Set up the initial 'ping' state of the ping-pong buffers */ qtmux->reserved_moov_first_active = TRUE; gst_qt_mux_configure_moov (qtmux); gst_qt_mux_setup_metadata (qtmux); /* Empty free atom to begin, starting on an 8-byte boundary */ ret = gst_qt_mux_send_free_atom (qtmux, &qtmux->header_size, 8, FALSE); if (ret != GST_FLOW_OK) return ret; /* Moov header, not padded yet */ ret = gst_qt_mux_send_moov (qtmux, &qtmux->header_size, 0, FALSE, FALSE); if (ret != GST_FLOW_OK) return ret; /* The moov we just sent contains the 'base' size of the moov, before * we put in any time-dependent per-trak data. Use that to make * a good estimate of how much extra to reserve */ /* Calculate how much space to reserve for our MOOV atom. * We actually reserve twice that, for ping-pong buffers */ qtmux->base_moov_size = qtmux->last_moov_size; GST_LOG_OBJECT (qtmux, "Base moov size is %u before any indexes", qtmux->base_moov_size); qtmux->reserved_moov_size = qtmux->base_moov_size + gst_util_uint64_scale (reserved_max_duration, reserved_bytes_per_sec_per_trak * atom_moov_get_trak_count (qtmux->moov), GST_SECOND); /* Need space for at least 4 atom headers. More really, but * this as an absolute minimum */ if (qtmux->reserved_moov_size < 4 * 8) goto reserved_moov_too_small; GST_DEBUG_OBJECT (qtmux, "reserving header area of size %u", 2 * qtmux->reserved_moov_size + 16); GST_OBJECT_LOCK (qtmux); qtmux->reserved_duration_remaining = gst_util_uint64_scale (qtmux->reserved_moov_size - qtmux->base_moov_size, GST_SECOND, reserved_bytes_per_sec_per_trak * atom_moov_get_trak_count (qtmux->moov)); GST_OBJECT_UNLOCK (qtmux); /* Now that we know how much reserved space is targetted, * output a free atom to fill the extra reserved */ ret = gst_qt_mux_send_free_atom (qtmux, &qtmux->header_size, qtmux->reserved_moov_size - qtmux->base_moov_size, FALSE); if (ret != GST_FLOW_OK) return ret; /* Then a free atom containing 'pong' buffer, with an * extra 8 bytes to account for the free atom header itself */ ret = gst_qt_mux_send_free_atom (qtmux, &qtmux->header_size, qtmux->reserved_moov_size + 8, FALSE); if (ret != GST_FLOW_OK) return ret; /* extra atoms go after the free/moov(s), before the mdat */ ret = gst_qt_mux_send_extra_atoms (qtmux, TRUE, &qtmux->header_size, FALSE); if (ret != GST_FLOW_OK) return ret; qtmux->mdat_pos = qtmux->header_size; /* extended atom in case we go over 4GB while writing and need * the full 64-bit atom */ ret = gst_qt_mux_send_mdat_header (qtmux, &qtmux->header_size, 0, TRUE, FALSE); break; case GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL: ret = gst_qt_mux_prepare_and_send_ftyp (qtmux); if (ret != GST_FLOW_OK) break; /* Store this as the moov offset for later updating. * We record mdat position below */ qtmux->moov_pos = qtmux->header_size; if (!gst_qt_mux_prefill_samples (qtmux)) { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, ("Unsupported codecs or configuration for prefill mode"), (NULL)); return GST_FLOW_ERROR; } gst_qt_mux_update_global_statistics (qtmux); gst_qt_mux_configure_moov (qtmux); gst_qt_mux_update_edit_lists (qtmux); gst_qt_mux_setup_metadata (qtmux); /* Moov header with pre-filled samples */ ret = gst_qt_mux_send_moov (qtmux, &qtmux->header_size, 0, FALSE, FALSE); if (ret != GST_FLOW_OK) return ret; /* last_moov_size now contains the full size of the moov, moov_pos the * position. This allows us to rewrite it in the very end as needed */ qtmux->reserved_moov_size = qtmux->last_moov_size + 12 * g_slist_length (qtmux->sinkpads) + 8; /* Send an additional free atom at the end so we definitely have space * to rewrite the moov header at the end and remove the samples that * were not actually written */ ret = gst_qt_mux_send_free_atom (qtmux, &qtmux->header_size, 12 * g_slist_length (qtmux->sinkpads) + 8, FALSE); if (ret != GST_FLOW_OK) return ret; /* extra atoms go after the free/moov(s), before the mdat */ ret = gst_qt_mux_send_extra_atoms (qtmux, TRUE, &qtmux->header_size, FALSE); if (ret != GST_FLOW_OK) return ret; qtmux->mdat_pos = qtmux->header_size; /* And now send the mdat header */ ret = gst_qt_mux_send_mdat_header (qtmux, &qtmux->header_size, qtmux->mdat_size, TRUE, FALSE); /* chunks position is set relative to the first byte of the * MDAT atom payload. Set the overall offset into the file */ atom_moov_chunks_set_offset (qtmux->moov, qtmux->header_size); { GstSegment segment; gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = qtmux->moov_pos; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); ret = gst_qt_mux_send_moov (qtmux, NULL, 0, FALSE, FALSE); if (ret != GST_FLOW_OK) return ret; segment.start = qtmux->header_size; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); } qtmux->current_chunk_size = 0; qtmux->current_chunk_duration = 0; qtmux->current_chunk_offset = -1; qtmux->mdat_size = 0; qtmux->current_pad = NULL; qtmux->longest_chunk = GST_CLOCK_TIME_NONE; for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; qtpad->total_bytes = 0; qtpad->total_duration = 0; qtpad->first_dts = qtpad->first_ts = GST_CLOCK_TIME_NONE; qtpad->last_dts = GST_CLOCK_TIME_NONE; qtpad->sample_offset = 0; } break; case GST_QT_MUX_MODE_FAST_START: GST_OBJECT_LOCK (qtmux); qtmux->fast_start_file = g_fopen (qtmux->fast_start_file_path, "wb+"); if (!qtmux->fast_start_file) goto open_failed; GST_OBJECT_UNLOCK (qtmux); /* send a dummy buffer for preroll */ ret = gst_qt_mux_send_buffer (qtmux, gst_buffer_new (), NULL, FALSE); break; case GST_QT_MUX_MODE_FRAGMENTED: case GST_QT_MUX_MODE_FRAGMENTED_STREAMABLE: ret = gst_qt_mux_prepare_and_send_ftyp (qtmux); if (ret != GST_FLOW_OK) break; /* store the moov pos so we can update the duration later * in non-streamable mode */ qtmux->moov_pos = qtmux->header_size; GST_DEBUG_OBJECT (qtmux, "fragment duration %d ms, writing headers", qtmux->fragment_duration); /* also used as snapshot marker to indicate fragmented file */ qtmux->fragment_sequence = 1; /* prepare moov and/or tags */ gst_qt_mux_configure_moov (qtmux); gst_qt_mux_setup_metadata (qtmux); ret = gst_qt_mux_send_moov (qtmux, &qtmux->header_size, 0, FALSE, FALSE); if (ret != GST_FLOW_OK) return ret; /* extra atoms */ ret = gst_qt_mux_send_extra_atoms (qtmux, TRUE, &qtmux->header_size, FALSE); if (ret != GST_FLOW_OK) break; /* prepare index if not streamable */ if (qtmux->mux_mode == GST_QT_MUX_MODE_FRAGMENTED) qtmux->mfra = atom_mfra_new (qtmux->context); break; } return ret; /* ERRORS */ invalid_isml: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, ("Cannot create an ISML file with 0 fragment duration"), (NULL)); return GST_FLOW_ERROR; } reserved_moov_too_small: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, ("Not enough reserved space for creating headers"), (NULL)); return GST_FLOW_ERROR; } open_failed: { GST_ELEMENT_ERROR (qtmux, RESOURCE, OPEN_READ_WRITE, (("Could not open temporary file \"%s\""), qtmux->fast_start_file_path), GST_ERROR_SYSTEM); GST_OBJECT_UNLOCK (qtmux); return GST_FLOW_ERROR; } } static GstFlowReturn gst_qt_mux_send_last_buffers (GstQTMux * qtmux) { GstFlowReturn ret = GST_FLOW_OK; GSList *walk; for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; /* avoid add_buffer complaining if not negotiated * in which case no buffers either, so skipping */ if (!qtpad->fourcc) { GST_DEBUG_OBJECT (qtmux, "Pad %s has never had buffers", GST_PAD_NAME (qtpad->collect.pad)); continue; } /* send last buffer; also flushes possibly queued buffers/ts */ GST_DEBUG_OBJECT (qtmux, "Sending the last buffer for pad %s", GST_PAD_NAME (qtpad->collect.pad)); ret = gst_qt_mux_add_buffer (qtmux, qtpad, NULL); if (ret != GST_FLOW_OK) { GST_WARNING_OBJECT (qtmux, "Failed to send last buffer for %s, " "flow return: %s", GST_PAD_NAME (qtpad->collect.pad), gst_flow_get_name (ret)); } } return ret; } static void gst_qt_mux_update_global_statistics (GstQTMux * qtmux) { GSList *walk; /* for setting some subtitles fields */ guint max_width = 0; guint max_height = 0; qtmux->first_ts = qtmux->last_dts = GST_CLOCK_TIME_NONE; for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; if (!qtpad->fourcc) { GST_DEBUG_OBJECT (qtmux, "Pad %s has never had buffers", GST_PAD_NAME (qtpad->collect.pad)); continue; } /* having flushed above, can check for buffers now */ if (GST_CLOCK_TIME_IS_VALID (qtpad->first_ts)) { GstClockTime first_pts_in = qtpad->first_ts; /* it should be, since we got first_ts by adding adjustment * to a positive incoming PTS */ if (qtpad->dts_adjustment <= first_pts_in) first_pts_in -= qtpad->dts_adjustment; /* determine max stream duration */ if (!GST_CLOCK_TIME_IS_VALID (qtmux->last_dts) || qtpad->last_dts > qtmux->last_dts) { qtmux->last_dts = qtpad->last_dts; } if (!GST_CLOCK_TIME_IS_VALID (qtmux->first_ts) || first_pts_in < qtmux->first_ts) { /* we need the original incoming PTS here, as this first_ts * is used in update_edit_lists to construct the edit list that arrange * for sync'ed streams. The first_ts is most likely obtained from * some (audio) stream with 0 dts_adjustment and initial 0 PTS, * so it makes no difference, though it matters in other cases */ qtmux->first_ts = first_pts_in; } } /* subtitles need to know the video width/height, * it is stored shifted 16 bits to the left according to the * spec */ max_width = MAX (max_width, (qtpad->trak->tkhd.width >> 16)); max_height = MAX (max_height, (qtpad->trak->tkhd.height >> 16)); /* update average bitrate of streams if needed */ { guint32 avgbitrate = 0; guint32 maxbitrate = qtpad->max_bitrate; if (qtpad->avg_bitrate) avgbitrate = qtpad->avg_bitrate; else if (qtpad->total_duration > 0) avgbitrate = (guint32) gst_util_uint64_scale_round (qtpad->total_bytes, 8 * GST_SECOND, qtpad->total_duration); atom_trak_update_bitrates (qtpad->trak, avgbitrate, maxbitrate); } } /* need to update values on subtitle traks now that we know the * max width and height */ for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; if (!qtpad->fourcc) { GST_DEBUG_OBJECT (qtmux, "Pad %s has never had buffers", GST_PAD_NAME (qtpad->collect.pad)); continue; } if (qtpad->fourcc == FOURCC_tx3g) { atom_trak_tx3g_update_dimension (qtpad->trak, max_width, max_height); } } } /* Called after gst_qt_mux_update_global_statistics() updates the * first_ts tracking, to create/set edit lists for delayed streams */ static void gst_qt_mux_update_edit_lists (GstQTMux * qtmux) { GSList *walk; GST_DEBUG_OBJECT (qtmux, "Media first ts selected: %" GST_TIME_FORMAT, GST_TIME_ARGS (qtmux->first_ts)); /* add/update EDTSs for late streams. configure_moov will have * set the trak durations above by summing the sample tables, * here we extend that if needing to insert an empty segment */ for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; atom_trak_edts_clear (qtpad->trak); if (GST_CLOCK_TIME_IS_VALID (qtpad->first_ts)) { guint32 lateness = 0; guint32 duration = qtpad->trak->tkhd.duration; gboolean has_gap; has_gap = (qtpad->first_ts > (qtmux->first_ts + qtpad->dts_adjustment)); if (has_gap) { GstClockTime diff, trak_lateness; diff = qtpad->first_ts - (qtmux->first_ts + qtpad->dts_adjustment); lateness = gst_util_uint64_scale_round (diff, qtmux->timescale, GST_SECOND); /* Allow up to 1 trak timescale unit of lateness, Such a small * timestamp/duration can't be represented by the trak-specific parts * of the headers anyway, so it's irrelevantly small */ trak_lateness = gst_util_uint64_scale (diff, atom_trak_get_timescale (qtpad->trak), GST_SECOND); if (trak_lateness > 0 && diff > qtmux->start_gap_threshold) { GST_DEBUG_OBJECT (qtmux, "Pad %s is a late stream by %" GST_TIME_FORMAT, GST_PAD_NAME (qtpad->collect.pad), GST_TIME_ARGS (diff)); atom_trak_set_elst_entry (qtpad->trak, 0, lateness, (guint32) - 1, (guint32) (1 * 65536.0)); } } /* Always write an edit list for the whole track. In general this is not * necessary except for the case of having a gap or DTS adjustment but * it allows to give the whole track's duration in the usually more * accurate media timescale */ { GstClockTime ctts = 0; guint32 media_start; if (qtpad->first_ts > qtpad->first_dts) ctts = qtpad->first_ts - qtpad->first_dts; media_start = gst_util_uint64_scale_round (ctts, atom_trak_get_timescale (qtpad->trak), GST_SECOND); /* atom_trak_set_elst_entry() has a quirk - if the edit list * is empty because there's no gap added above, this call * will not replace index 1, it will create the entry at index 0. * Luckily, that's exactly what we want here */ atom_trak_set_elst_entry (qtpad->trak, 1, duration, media_start, (guint32) (1 * 65536.0)); } /* need to add the empty time to the trak duration */ duration += lateness; qtpad->trak->tkhd.duration = duration; if (qtpad->tc_trak) { qtpad->tc_trak->tkhd.duration = duration; qtpad->tc_trak->mdia.mdhd.time_info.duration = duration; } /* And possibly grow the moov duration */ if (duration > qtmux->moov->mvhd.time_info.duration) { qtmux->moov->mvhd.time_info.duration = duration; qtmux->moov->mvex.mehd.fragment_duration = duration; } } } } static GstFlowReturn gst_qt_mux_update_timecode (GstQTMux * qtmux, GstQTPad * qtpad) { GstSegment segment; GstBuffer *buf; GstMapInfo map; guint64 offset = qtpad->tc_pos; GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); if (qtmux_klass->format != GST_QT_MUX_FORMAT_QT) return GST_FLOW_OK; g_assert (qtpad->tc_pos != -1); gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = offset; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); buf = gst_buffer_new_and_alloc (4); gst_buffer_map (buf, &map, GST_MAP_WRITE); GST_WRITE_UINT32_BE (map.data, gst_video_time_code_frames_since_daily_jam (qtpad->first_tc)); gst_buffer_unmap (buf, &map); /* Reset this value, so the timecode won't be re-rewritten */ qtpad->tc_pos = -1; return gst_qt_mux_send_buffer (qtmux, buf, &offset, FALSE); } static GstFlowReturn gst_qt_mux_stop_file (GstQTMux * qtmux) { gboolean ret = GST_FLOW_OK; guint64 offset = 0, size = 0; gboolean large_file; GSList *walk; GST_DEBUG_OBJECT (qtmux, "Updating remaining values and sending last data"); /* pushing last buffers for each pad */ if ((ret = gst_qt_mux_send_last_buffers (qtmux)) != GST_FLOW_OK) return ret; if (qtmux->mux_mode == GST_QT_MUX_MODE_FRAGMENTED_STREAMABLE) { /* Streamable mode; no need to write duration or MFRA */ GST_DEBUG_OBJECT (qtmux, "streamable file; nothing to stop"); return GST_FLOW_OK; } gst_qt_mux_update_global_statistics (qtmux); for (walk = qtmux->collect->data; walk; walk = walk->next) { GstQTPad *qtpad = (GstQTPad *) walk->data; if (qtpad->tc_pos != -1) { /* File is being stopped and timecode hasn't been updated. Update it now * with whatever we have */ ret = gst_qt_mux_update_timecode (qtmux, qtpad); if (ret != GST_FLOW_OK) return ret; } } switch (qtmux->mux_mode) { case GST_QT_MUX_MODE_FRAGMENTED:{ GstSegment segment; guint8 *data = NULL; GstBuffer *buf; size = offset = 0; GST_DEBUG_OBJECT (qtmux, "adding mfra"); if (!atom_mfra_copy_data (qtmux->mfra, &data, &size, &offset)) goto serialize_error; buf = _gst_buffer_new_take_data (data, offset); ret = gst_qt_mux_send_buffer (qtmux, buf, NULL, FALSE); if (ret != GST_FLOW_OK) return ret; /* only mvex duration is updated, * mvhd should be consistent with empty moov * (but TODO maybe some clients do not handle that well ?) */ qtmux->moov->mvex.mehd.fragment_duration = gst_util_uint64_scale_round (qtmux->last_dts, qtmux->timescale, GST_SECOND); GST_DEBUG_OBJECT (qtmux, "rewriting moov with mvex duration %" GST_TIME_FORMAT, GST_TIME_ARGS (qtmux->last_dts)); /* seek and rewrite the header */ gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = qtmux->moov_pos; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); /* no need to seek back */ return gst_qt_mux_send_moov (qtmux, NULL, 0, FALSE, FALSE); } case GST_QT_MUX_MODE_ROBUST_RECORDING:{ ret = gst_qt_mux_robust_recording_rewrite_moov (qtmux); if (G_UNLIKELY (ret != GST_FLOW_OK)) return ret; /* Finalise by writing the final size into the mdat. Up until now * it's been 0, which means 'rest of the file' * No need to seek back after this, we won't write any more */ return gst_qt_mux_update_mdat_size (qtmux, qtmux->mdat_pos, qtmux->mdat_size, NULL, TRUE); } case GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL:{ GSList *walk; guint32 next_track_id = qtmux->moov->mvhd.next_track_id; for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qpad = (GstQTPad *) cdata; guint64 block_idx; AtomSTBL *stbl = &qpad->trak->mdia.minf.stbl; /* Get the block index of the last sample we wrote, not of the next * sample we would write */ block_idx = prefill_get_block_index (qtmux, qpad); /* stts */ if (block_idx > 0) { STTSEntry *entry; guint64 nsamples = 0; gint i, n; n = atom_array_get_len (&stbl->stts.entries); for (i = 0; i < n; i++) { entry = &atom_array_index (&stbl->stts.entries, i); if (nsamples + entry->sample_count >= qpad->sample_offset) { entry->sample_count = qpad->sample_offset - nsamples; stbl->stts.entries.len = i + 1; break; } nsamples += entry->sample_count; } g_assert (i < n); } else { stbl->stts.entries.len = 0; } /* stsz */ { g_assert (stbl->stsz.entries.len == 0); stbl->stsz.table_size = qpad->sample_offset; } /* stco/stsc */ { gint i, n; guint64 nsamples = 0; gint chunk_index = 0; const TrakBufferEntryInfo *sample_entry; if (block_idx > 0) { sample_entry = &g_array_index (qpad->samples, TrakBufferEntryInfo, block_idx - 1); n = stbl->stco64.entries.len; for (i = 0; i < n; i++) { guint64 *entry = &atom_array_index (&stbl->stco64.entries, i); if (*entry == sample_entry->chunk_offset) { stbl->stco64.entries.len = i + 1; chunk_index = i + 1; break; } } g_assert (i < n); g_assert (chunk_index > 0); n = stbl->stsc.entries.len; for (i = 0; i < n; i++) { STSCEntry *entry = &atom_array_index (&stbl->stsc.entries, i); if (entry->first_chunk >= chunk_index) break; if (i > 0) { nsamples += (entry->first_chunk - atom_array_index (&stbl->stsc.entries, i - 1).first_chunk) * atom_array_index (&stbl->stsc.entries, i - 1).samples_per_chunk; } } g_assert (i <= n); if (i > 0) { STSCEntry *prev_entry = &atom_array_index (&stbl->stsc.entries, i - 1); nsamples += (chunk_index - prev_entry->first_chunk) * prev_entry->samples_per_chunk; if (qpad->sample_offset - nsamples > 0) { stbl->stsc.entries.len = i; atom_stsc_add_new_entry (&stbl->stsc, chunk_index, qpad->sample_offset - nsamples); } else { stbl->stsc.entries.len = i; stbl->stco64.entries.len--; } } else { /* Everything in a single chunk */ stbl->stsc.entries.len = 0; atom_stsc_add_new_entry (&stbl->stsc, chunk_index, qpad->sample_offset); } } else { stbl->stco64.entries.len = 0; stbl->stsc.entries.len = 0; } } { GList *walk2; for (walk2 = qtmux->moov->mvex.trexs; walk2; walk2 = walk2->next) { AtomTREX *trex = walk2->data; if (trex->track_ID == qpad->trak->tkhd.track_ID) { trex->track_ID = next_track_id; break; } } qpad->trak->tkhd.track_ID = next_track_id++; } } qtmux->moov->mvhd.next_track_id = next_track_id; gst_qt_mux_update_global_statistics (qtmux); gst_qt_mux_configure_moov (qtmux); gst_qt_mux_update_edit_lists (qtmux); /* Check if any gap edit lists were added. We don't have any space * reserved for this in the moov and the pre-finalized moov would have * broken A/V synchronization. Error out here now */ for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qpad = (GstQTPad *) cdata; if (qpad->trak->edts && g_slist_length (qpad->trak->edts->elst.entries) > 1) { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Can't support gaps in prefill mode")); return GST_FLOW_ERROR; } } gst_qt_mux_setup_metadata (qtmux); atom_moov_chunks_set_offset (qtmux->moov, qtmux->header_size); { GstSegment segment; gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = qtmux->moov_pos; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); ret = gst_qt_mux_send_moov (qtmux, NULL, qtmux->reserved_moov_size, FALSE, FALSE); if (ret != GST_FLOW_OK) return ret; if (qtmux->reserved_moov_size > qtmux->last_moov_size) { ret = gst_qt_mux_send_free_atom (qtmux, NULL, qtmux->reserved_moov_size - qtmux->last_moov_size, TRUE); } if (ret != GST_FLOW_OK) return ret; } ret = gst_qt_mux_update_mdat_size (qtmux, qtmux->mdat_pos, qtmux->mdat_size, NULL, FALSE); return ret; } default: break; } /* Moov-at-end or fast-start mode from here down */ gst_qt_mux_configure_moov (qtmux); gst_qt_mux_update_edit_lists (qtmux); /* tags into file metadata */ gst_qt_mux_setup_metadata (qtmux); large_file = (qtmux->mdat_size > MDAT_LARGE_FILE_LIMIT); switch (qtmux->mux_mode) { case GST_QT_MUX_MODE_FAST_START:{ /* if faststart, update the offset of the atoms in the movie with the offset * that the movie headers before mdat will cause. * Also, send the ftyp */ offset = size = 0; ret = gst_qt_mux_prepare_and_send_ftyp (qtmux); if (ret != GST_FLOW_OK) { goto ftyp_error; } /* copy into NULL to obtain size */ if (!atom_moov_copy_data (qtmux->moov, NULL, &size, &offset)) goto serialize_error; GST_DEBUG_OBJECT (qtmux, "calculated moov atom size %" G_GUINT64_FORMAT, offset); offset += qtmux->header_size + (large_file ? 16 : 8); /* sum up with the extra atoms size */ ret = gst_qt_mux_send_extra_atoms (qtmux, FALSE, &offset, FALSE); if (ret != GST_FLOW_OK) return ret; break; } default: offset = qtmux->header_size; break; } /* Now that we know the size of moov + extra atoms, we can adjust * the chunk offsets stored into the moov */ atom_moov_chunks_set_offset (qtmux->moov, offset); /* write out moov and extra atoms */ /* note: as of this point, we no longer care about tracking written data size, * since there is no more use for it anyway */ ret = gst_qt_mux_send_moov (qtmux, NULL, 0, FALSE, FALSE); if (ret != GST_FLOW_OK) return ret; /* extra atoms */ ret = gst_qt_mux_send_extra_atoms (qtmux, TRUE, NULL, FALSE); if (ret != GST_FLOW_OK) return ret; switch (qtmux->mux_mode) { case GST_QT_MUX_MODE_MOOV_AT_END: { /* mdat needs update iff not using faststart */ GST_DEBUG_OBJECT (qtmux, "updating mdat size"); ret = gst_qt_mux_update_mdat_size (qtmux, qtmux->mdat_pos, qtmux->mdat_size, NULL, FALSE); /* note; no seeking back to the end of file is done, * since we no longer write anything anyway */ break; } case GST_QT_MUX_MODE_FAST_START: { /* send mdat atom and move buffered data into it */ /* mdat_size = accumulated (buffered data) */ ret = gst_qt_mux_send_mdat_header (qtmux, NULL, qtmux->mdat_size, large_file, FALSE); if (ret != GST_FLOW_OK) return ret; ret = gst_qt_mux_send_buffered_data (qtmux, NULL); if (ret != GST_FLOW_OK) return ret; break; } default: g_assert_not_reached (); } return ret; /* ERRORS */ serialize_error: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Failed to serialize moov")); return GST_FLOW_ERROR; } ftyp_error: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Failed to send ftyp")); return GST_FLOW_ERROR; } } static GstFlowReturn gst_qt_mux_pad_fragment_add_buffer (GstQTMux * qtmux, GstQTPad * pad, GstBuffer * buf, gboolean force, guint32 nsamples, gint64 dts, guint32 delta, guint32 size, gboolean sync, gint64 pts_offset) { GstFlowReturn ret = GST_FLOW_OK; /* setup if needed */ if (G_UNLIKELY (!pad->traf || force)) goto init; flush: /* flush pad fragment if threshold reached, * or at new keyframe if we should be minding those in the first place */ if (G_UNLIKELY (force || (sync && pad->sync) || pad->fragment_duration < (gint64) delta)) { AtomMOOF *moof; guint64 size = 0, offset = 0; guint8 *data = NULL; GstBuffer *buffer; guint i, total_size; /* now we know where moof ends up, update offset in tfra */ if (pad->tfra) atom_tfra_update_offset (pad->tfra, qtmux->header_size); moof = atom_moof_new (qtmux->context, qtmux->fragment_sequence); /* takes ownership */ atom_moof_add_traf (moof, pad->traf); pad->traf = NULL; atom_moof_copy_data (moof, &data, &size, &offset); buffer = _gst_buffer_new_take_data (data, offset); GST_LOG_OBJECT (qtmux, "writing moof size %" G_GSIZE_FORMAT, gst_buffer_get_size (buffer)); ret = gst_qt_mux_send_buffer (qtmux, buffer, &qtmux->header_size, FALSE); /* and actual data */ total_size = 0; for (i = 0; i < atom_array_get_len (&pad->fragment_buffers); i++) { total_size += gst_buffer_get_size (atom_array_index (&pad->fragment_buffers, i)); } GST_LOG_OBJECT (qtmux, "writing %d buffers, total_size %d", atom_array_get_len (&pad->fragment_buffers), total_size); if (ret == GST_FLOW_OK) ret = gst_qt_mux_send_mdat_header (qtmux, &qtmux->header_size, total_size, FALSE, FALSE); for (i = 0; i < atom_array_get_len (&pad->fragment_buffers); i++) { if (G_LIKELY (ret == GST_FLOW_OK)) ret = gst_qt_mux_send_buffer (qtmux, atom_array_index (&pad->fragment_buffers, i), &qtmux->header_size, FALSE); else gst_buffer_unref (atom_array_index (&pad->fragment_buffers, i)); } atom_array_clear (&pad->fragment_buffers); atom_moof_free (moof); qtmux->fragment_sequence++; force = FALSE; } init: if (G_UNLIKELY (!pad->traf)) { GST_LOG_OBJECT (qtmux, "setting up new fragment"); pad->traf = atom_traf_new (qtmux->context, atom_trak_get_id (pad->trak)); atom_array_init (&pad->fragment_buffers, 512); pad->fragment_duration = gst_util_uint64_scale (qtmux->fragment_duration, atom_trak_get_timescale (pad->trak), 1000); if (G_UNLIKELY (qtmux->mfra && !pad->tfra)) { pad->tfra = atom_tfra_new (qtmux->context, atom_trak_get_id (pad->trak)); atom_mfra_add_tfra (qtmux->mfra, pad->tfra); } atom_traf_set_base_decode_time (pad->traf, dts); } /* add buffer and metadata */ atom_traf_add_samples (pad->traf, delta, size, sync, pts_offset, pad->sync && sync); atom_array_append (&pad->fragment_buffers, buf, 256); pad->fragment_duration -= delta; if (pad->tfra) { guint32 sn = atom_traf_get_sample_num (pad->traf); if ((sync && pad->sync) || (sn == 1 && !pad->sync)) atom_tfra_add_entry (pad->tfra, dts, sn); } if (G_UNLIKELY (force)) goto flush; return ret; } /* Here's the clever bit of robust recording: Updating the moov * header is done using a ping-pong scheme inside 2 blocks of size * 'reserved_moov_size' at the start of the file, in such a way that the * file on-disk is always valid if interrupted. * Inside the reserved space, we have 2 pairs of free + moov atoms * (in that order), free-A + moov-A @ offset 0 and free-B + moov-B at * at offset "reserved_moov_size". * * 1. Free-A has 0 size payload, moov-A immediately after is * active/current, and is padded with an internal Free atom to * end at reserved_space/2. Free-B is at reserved_space/2, sized * to cover the remaining free space (including moov-B). * 2. We write moov-B (which is invisible inside free-B), and pad it to * end at the end of free space. Then, we update free-A to size * reserved_space/2 + sizeof(free-B), which hides moov-A and the * free-B header, and makes moov-B active. * 3. Rewrite moov-A inside free-A, with padding out to free-B. * Change the size of free-A to make moov-A active again. * 4. Rinse and repeat. * */ static GstFlowReturn gst_qt_mux_robust_recording_rewrite_moov (GstQTMux * qtmux) { GstSegment segment; GstFlowReturn ret; guint64 freeA_offset; guint32 new_freeA_size; guint64 new_moov_offset; /* Update moov info, then seek and rewrite the MOOV atom */ gst_qt_mux_update_global_statistics (qtmux); gst_qt_mux_configure_moov (qtmux); gst_qt_mux_update_edit_lists (qtmux); /* tags into file metadata */ gst_qt_mux_setup_metadata (qtmux); /* chunks position is set relative to the first byte of the * MDAT atom payload. Set the overall offset into the file */ atom_moov_chunks_set_offset (qtmux->moov, qtmux->header_size); /* Calculate which moov to rewrite. qtmux->moov_pos points to * the start of the free-A header */ freeA_offset = qtmux->moov_pos; if (qtmux->reserved_moov_first_active) { GST_DEBUG_OBJECT (qtmux, "Updating pong moov header"); /* After this, freeA will include itself, moovA, plus the freeB * header */ new_freeA_size = qtmux->reserved_moov_size + 16; } else { GST_DEBUG_OBJECT (qtmux, "Updating ping moov header"); new_freeA_size = 8; } /* the moov we update is after free-A, calculate its offset */ new_moov_offset = freeA_offset + new_freeA_size; /* Swap ping-pong cadence marker */ qtmux->reserved_moov_first_active = !qtmux->reserved_moov_first_active; /* seek and rewrite the MOOV atom */ gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = new_moov_offset; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); ret = gst_qt_mux_send_moov (qtmux, NULL, qtmux->reserved_moov_size, FALSE, TRUE); if (ret != GST_FLOW_OK) return ret; /* Update the estimated recording space remaining, based on amount used so * far and duration muxed so far */ if (qtmux->last_moov_size > qtmux->base_moov_size && qtmux->last_dts > 0) { GstClockTime remain; GstClockTime time_muxed = qtmux->last_dts; remain = gst_util_uint64_scale (qtmux->reserved_moov_size - qtmux->last_moov_size, time_muxed, qtmux->last_moov_size - qtmux->base_moov_size); /* Always under-estimate slightly, so users * have time to stop muxing before we run out */ if (remain < GST_SECOND / 2) remain = 0; else remain -= GST_SECOND / 2; GST_INFO_OBJECT (qtmux, "Reserved %u header bytes. Used %u in %" GST_TIME_FORMAT ". Remaining now %u or approx %" G_GUINT64_FORMAT " ns\n", qtmux->reserved_moov_size, qtmux->last_moov_size, GST_TIME_ARGS (qtmux->last_dts), qtmux->reserved_moov_size - qtmux->last_moov_size, remain); GST_OBJECT_LOCK (qtmux); qtmux->reserved_duration_remaining = remain; qtmux->muxed_since_last_update = 0; GST_DEBUG_OBJECT (qtmux, "reserved remaining duration now %" G_GUINT64_FORMAT, qtmux->reserved_duration_remaining); GST_OBJECT_UNLOCK (qtmux); } /* Now update the moov-A size. Don't pass offset, since we don't need * send_free_atom() to seek for us - all our callers seek back to * where they need after this, or they don't need it */ gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = freeA_offset; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); ret = gst_qt_mux_send_free_atom (qtmux, NULL, new_freeA_size, TRUE); return ret; } static GstFlowReturn gst_qt_mux_robust_recording_update (GstQTMux * qtmux, GstClockTime position) { GstSegment segment; GstFlowReturn flow_ret; guint64 mdat_offset = qtmux->mdat_pos + 16 + qtmux->mdat_size; GST_OBJECT_LOCK (qtmux); /* Update the offset of how much we've muxed, so the * report of remaining space keeps counting down */ if (position > qtmux->last_moov_update && position - qtmux->last_moov_update > qtmux->muxed_since_last_update) { GST_LOG_OBJECT (qtmux, "Muxed time %" G_GUINT64_FORMAT " since last moov update", qtmux->muxed_since_last_update); qtmux->muxed_since_last_update = position - qtmux->last_moov_update; } /* Next, check if we're supposed to send periodic moov updates downstream */ if (qtmux->reserved_moov_update_period == GST_CLOCK_TIME_NONE) { GST_OBJECT_UNLOCK (qtmux); return GST_FLOW_OK; } /* Update if position is > the threshold or there's been no update yet */ if (qtmux->last_moov_update != GST_CLOCK_TIME_NONE && (position <= qtmux->last_moov_update || (position - qtmux->last_moov_update) < qtmux->reserved_moov_update_period)) { GST_OBJECT_UNLOCK (qtmux); return GST_FLOW_OK; /* No update needed yet */ } qtmux->last_moov_update = position; GST_OBJECT_UNLOCK (qtmux); GST_DEBUG_OBJECT (qtmux, "Update moov atom, position %" GST_TIME_FORMAT " mdat starts @ %" G_GUINT64_FORMAT " we were a %" G_GUINT64_FORMAT, GST_TIME_ARGS (position), qtmux->mdat_pos, mdat_offset); flow_ret = gst_qt_mux_robust_recording_rewrite_moov (qtmux); if (G_UNLIKELY (flow_ret != GST_FLOW_OK)) return flow_ret; /* Seek back to previous position */ gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = mdat_offset; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); return flow_ret; } static GstFlowReturn gst_qt_mux_register_and_push_sample (GstQTMux * qtmux, GstQTPad * pad, GstBuffer * buffer, gboolean is_last_buffer, guint nsamples, gint64 last_dts, gint64 scaled_duration, guint sample_size, guint64 chunk_offset, gboolean sync, gboolean do_pts, gint64 pts_offset) { GstFlowReturn ret = GST_FLOW_OK; /* note that a new chunk is started each time (not fancy but works) */ if (qtmux->moov_recov_file) { if (!atoms_recov_write_trak_samples (qtmux->moov_recov_file, pad->trak, nsamples, (gint32) scaled_duration, sample_size, chunk_offset, sync, do_pts, pts_offset)) { GST_WARNING_OBJECT (qtmux, "Failed to write sample information to " "recovery file, disabling recovery"); fclose (qtmux->moov_recov_file); qtmux->moov_recov_file = NULL; } } switch (qtmux->mux_mode) { case GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL:{ const TrakBufferEntryInfo *sample_entry; guint64 block_idx = prefill_get_block_index (qtmux, pad); if (block_idx >= pad->samples->len) { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Unexpected sample %" G_GUINT64_FORMAT ", expected up to %u", block_idx, pad->samples->len)); gst_buffer_unref (buffer); return GST_FLOW_ERROR; } /* Check if all values are as expected */ sample_entry = &g_array_index (pad->samples, TrakBufferEntryInfo, block_idx); /* Allow +/- 1 difference for the scaled_duration to allow * for some rounding errors */ if (sample_entry->nsamples != nsamples || ABSDIFF (sample_entry->delta, scaled_duration) > 1 || sample_entry->size != sample_size || sample_entry->chunk_offset != chunk_offset || sample_entry->pts_offset != pts_offset || sample_entry->sync != sync) { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Unexpected values in sample %" G_GUINT64_FORMAT, pad->sample_offset + 1)); GST_ERROR_OBJECT (qtmux, "Expected: samples %u, delta %u, size %u, " "chunk offset %" G_GUINT64_FORMAT ", " "pts offset %" G_GUINT64_FORMAT ", sync %d", sample_entry->nsamples, sample_entry->delta, sample_entry->size, sample_entry->chunk_offset, sample_entry->pts_offset, sample_entry->sync); GST_ERROR_OBJECT (qtmux, "Got: samples %u, delta %u, size %u, " "chunk offset %" G_GUINT64_FORMAT ", " "pts offset %" G_GUINT64_FORMAT ", sync %d", nsamples, (guint) scaled_duration, sample_size, chunk_offset, pts_offset, sync); gst_buffer_unref (buffer); return GST_FLOW_ERROR; } ret = gst_qt_mux_send_buffer (qtmux, buffer, &qtmux->mdat_size, TRUE); break; } case GST_QT_MUX_MODE_MOOV_AT_END: case GST_QT_MUX_MODE_FAST_START: case GST_QT_MUX_MODE_ROBUST_RECORDING: atom_trak_add_samples (pad->trak, nsamples, (gint32) scaled_duration, sample_size, chunk_offset, sync, pts_offset); ret = gst_qt_mux_send_buffer (qtmux, buffer, &qtmux->mdat_size, TRUE); /* Check if it's time to re-write the headers in robust-recording mode */ if (ret == GST_FLOW_OK && qtmux->mux_mode == GST_QT_MUX_MODE_ROBUST_RECORDING) ret = gst_qt_mux_robust_recording_update (qtmux, pad->total_duration); break; case GST_QT_MUX_MODE_FRAGMENTED: case GST_QT_MUX_MODE_FRAGMENTED_STREAMABLE: /* ensure that always sync samples are marked as such */ ret = gst_qt_mux_pad_fragment_add_buffer (qtmux, pad, buffer, is_last_buffer, nsamples, last_dts, (gint32) scaled_duration, sample_size, !pad->sync || sync, pts_offset); break; } return ret; } static void gst_qt_mux_register_buffer_in_chunk (GstQTMux * qtmux, GstQTPad * pad, guint buffer_size, GstClockTime duration) { /* not that much happens here, * but updating any of this very likely needs to happen all in sync, * unless there is a very good reason not to */ /* for computing the avg bitrate */ pad->total_bytes += buffer_size; pad->total_duration += duration; /* for keeping track of where we are in chunk; * ensures that data really is located as recorded in atoms */ qtmux->current_chunk_size += buffer_size; qtmux->current_chunk_duration += duration; } static GstFlowReturn gst_qt_mux_check_and_update_timecode (GstQTMux * qtmux, GstQTPad * pad, GstBuffer * buf, GstFlowReturn ret) { GstVideoTimeCodeMeta *tc_meta; GstVideoTimeCode *tc; GstBuffer *tc_buf; gsize szret; guint32 frames_since_daily_jam; GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); if (!pad->trak->is_video) return ret; if (qtmux_klass->format != GST_QT_MUX_FORMAT_QT) return ret; if (buf == NULL || (pad->tc_trak != NULL && pad->tc_pos == -1)) return ret; tc_meta = gst_buffer_get_video_time_code_meta (buf); if (!tc_meta) return ret; tc = &tc_meta->tc; /* This means we never got a timecode before */ if (pad->first_tc == NULL) { #ifndef GST_DISABLE_GST_DEBUG gchar *tc_str = gst_video_time_code_to_string (tc); GST_DEBUG_OBJECT (qtmux, "Found first timecode %s", tc_str); g_free (tc_str); #endif g_assert (pad->tc_trak == NULL); pad->first_tc = gst_video_time_code_copy (tc); /* If frames are out of order, the frame we're currently getting might * not be the first one. Just write a 0 timecode for now and wait * until we receive a timecode that's lower than the current one */ if (pad->is_out_of_order) { pad->first_pts = GST_BUFFER_PTS (buf); frames_since_daily_jam = 0; /* Position to rewrite */ pad->tc_pos = qtmux->mdat_size; } else { frames_since_daily_jam = gst_video_time_code_frames_since_daily_jam (pad->first_tc); frames_since_daily_jam = GUINT32_TO_BE (frames_since_daily_jam); } /* Write the timecode trak now */ pad->tc_trak = atom_trak_new (qtmux->context); atom_moov_add_trak (qtmux->moov, pad->tc_trak); pad->trak->tref = atom_tref_new (FOURCC_tmcd); atom_tref_add_entry (pad->trak->tref, pad->tc_trak->tkhd.track_ID); atom_trak_set_timecode_type (pad->tc_trak, qtmux->context, pad->trak->mdia.mdhd.time_info.timescale, pad->first_tc); tc_buf = gst_buffer_new_allocate (NULL, 4, NULL); szret = gst_buffer_fill (tc_buf, 0, &frames_since_daily_jam, 4); g_assert (szret == 4); atom_trak_add_samples (pad->tc_trak, 1, 1, 4, qtmux->mdat_size, FALSE, 0); ret = gst_qt_mux_send_buffer (qtmux, tc_buf, &qtmux->mdat_size, TRUE); /* Need to reset the current chunk (of the previous pad) here because * some other data was written now above, and the pad has to start a * new chunk now */ qtmux->current_chunk_offset = -1; qtmux->current_chunk_size = 0; qtmux->current_chunk_duration = 0; } else if (qtmux->mux_mode == GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL) { frames_since_daily_jam = gst_video_time_code_frames_since_daily_jam (pad->first_tc); frames_since_daily_jam = GUINT32_TO_BE (frames_since_daily_jam); tc_buf = gst_buffer_new_allocate (NULL, 4, NULL); szret = gst_buffer_fill (tc_buf, 0, &frames_since_daily_jam, 4); g_assert (szret == 4); ret = gst_qt_mux_send_buffer (qtmux, tc_buf, &qtmux->mdat_size, TRUE); pad->tc_pos = -1; qtmux->current_chunk_offset = -1; qtmux->current_chunk_size = 0; qtmux->current_chunk_duration = 0; } else if (pad->is_out_of_order) { /* Check for a lower timecode than the one stored */ g_assert (pad->tc_trak != NULL); if (GST_BUFFER_DTS (buf) <= pad->first_pts) { if (gst_video_time_code_compare (tc, pad->first_tc) == -1) { gst_video_time_code_free (pad->first_tc); pad->first_tc = gst_video_time_code_copy (tc); } } else { guint64 bk_size = qtmux->mdat_size; GstSegment segment; /* If this frame's DTS is after the first PTS received, it means * we've already received the first frame to be presented. Otherwise * the decoder would need to go back in time */ gst_qt_mux_update_timecode (qtmux, pad); /* Reset writing position */ gst_segment_init (&segment, GST_FORMAT_BYTES); segment.start = bk_size; gst_pad_push_event (qtmux->srcpad, gst_event_new_segment (&segment)); } } return ret; } /* * Here we push the buffer and update the tables in the track atoms */ static GstFlowReturn gst_qt_mux_add_buffer (GstQTMux * qtmux, GstQTPad * pad, GstBuffer * buf) { GstBuffer *last_buf = NULL; GstClockTime duration; guint nsamples, sample_size; guint64 chunk_offset; gint64 last_dts, scaled_duration; gint64 pts_offset = 0; gboolean sync = FALSE; GstFlowReturn ret = GST_FLOW_OK; guint buffer_size; if (!pad->fourcc) goto not_negotiated; /* if this pad has a prepare function, call it */ if (pad->prepare_buf_func != NULL) { GstBuffer *new_buf; new_buf = pad->prepare_buf_func (pad, buf, qtmux); if (buf && !new_buf) return GST_FLOW_OK; buf = new_buf; } ret = gst_qt_mux_check_and_update_timecode (qtmux, pad, buf, ret); if (ret != GST_FLOW_OK) { if (buf) gst_buffer_unref (buf); return ret; } last_buf = pad->last_buf; pad->last_buf = buf; if (last_buf == NULL) { #ifndef GST_DISABLE_GST_DEBUG if (buf == NULL) { GST_DEBUG_OBJECT (qtmux, "Pad %s has no previous buffer stored and " "received NULL buffer, doing nothing", GST_PAD_NAME (pad->collect.pad)); } else { GST_LOG_OBJECT (qtmux, "Pad %s has no previous buffer stored, storing now", GST_PAD_NAME (pad->collect.pad)); } #endif goto exit; } if (!GST_BUFFER_PTS_IS_VALID (last_buf)) goto no_pts; /* if this is the first buffer, store the timestamp */ if (G_UNLIKELY (pad->first_ts == GST_CLOCK_TIME_NONE)) { if (GST_BUFFER_PTS_IS_VALID (last_buf)) { pad->first_ts = GST_BUFFER_PTS (last_buf); } else if (GST_BUFFER_DTS_IS_VALID (last_buf)) { pad->first_ts = GST_BUFFER_DTS (last_buf); } if (GST_BUFFER_DTS_IS_VALID (last_buf)) { pad->first_dts = pad->last_dts = GST_BUFFER_DTS (last_buf); } else if (GST_BUFFER_PTS_IS_VALID (last_buf)) { pad->first_dts = pad->last_dts = GST_BUFFER_PTS (last_buf); } if (GST_CLOCK_TIME_IS_VALID (pad->first_ts)) { GST_DEBUG ("setting first_ts to %" G_GUINT64_FORMAT, pad->first_ts); } else { GST_WARNING_OBJECT (qtmux, "First buffer for pad %s has no timestamp, " "using 0 as first timestamp", GST_PAD_NAME (pad->collect.pad)); pad->first_ts = pad->first_dts = 0; } GST_DEBUG_OBJECT (qtmux, "Stored first timestamp for pad %s %" GST_TIME_FORMAT, GST_PAD_NAME (pad->collect.pad), GST_TIME_ARGS (pad->first_ts)); } if (buf && GST_CLOCK_TIME_IS_VALID (GST_BUFFER_DTS (buf)) && GST_CLOCK_TIME_IS_VALID (GST_BUFFER_DTS (last_buf)) && GST_BUFFER_DTS (buf) < GST_BUFFER_DTS (last_buf)) { GST_ERROR ("decreasing DTS value %" GST_TIME_FORMAT " < %" GST_TIME_FORMAT, GST_TIME_ARGS (GST_BUFFER_DTS (buf)), GST_TIME_ARGS (GST_BUFFER_DTS (last_buf))); pad->last_buf = buf = gst_buffer_make_writable (buf); GST_BUFFER_DTS (buf) = GST_BUFFER_DTS (last_buf); } buffer_size = gst_buffer_get_size (last_buf); if (qtmux->mux_mode == GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL) { guint required_buffer_size = prefill_get_sample_size (qtmux, pad); guint fill_size = required_buffer_size - buffer_size; GstMemory *mem; GstMapInfo map; if (required_buffer_size < buffer_size) { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Sample size %u bigger than expected maximum %u", buffer_size, required_buffer_size)); goto bail; } if (fill_size > 0) { GST_DEBUG_OBJECT (qtmux, "Padding buffer by %u bytes to reach required %u bytes", fill_size, required_buffer_size); mem = gst_allocator_alloc (NULL, fill_size, NULL); gst_memory_map (mem, &map, GST_MAP_WRITE); memset (map.data, 0, map.size); gst_memory_unmap (mem, &map); last_buf = gst_buffer_make_writable (last_buf); gst_buffer_append_memory (last_buf, mem); buffer_size = required_buffer_size; } } /* duration actually means time delta between samples, so we calculate * the duration based on the difference in DTS or PTS, falling back * to DURATION if the other two don't exist, such as with the last * sample before EOS. Or use 0 if nothing else is available */ if (GST_BUFFER_DURATION_IS_VALID (last_buf)) duration = GST_BUFFER_DURATION (last_buf); else duration = 0; if (!pad->sparse) { if (buf && GST_BUFFER_DTS_IS_VALID (buf) && GST_BUFFER_DTS_IS_VALID (last_buf)) duration = GST_BUFFER_DTS (buf) - GST_BUFFER_DTS (last_buf); else if (buf && GST_BUFFER_PTS_IS_VALID (buf) && GST_BUFFER_PTS_IS_VALID (last_buf)) duration = GST_BUFFER_PTS (buf) - GST_BUFFER_PTS (last_buf); } if (qtmux->current_pad != pad || qtmux->current_chunk_offset == -1) { GST_DEBUG_OBJECT (qtmux, "Switching to next chunk for pad %s:%s: offset %" G_GUINT64_FORMAT ", size %" G_GUINT64_FORMAT ", duration %" GST_TIME_FORMAT, GST_DEBUG_PAD_NAME (pad->collect.pad), qtmux->current_chunk_offset, qtmux->current_chunk_size, GST_TIME_ARGS (qtmux->current_chunk_duration)); qtmux->current_pad = pad; if (qtmux->current_chunk_offset == -1) qtmux->current_chunk_offset = qtmux->mdat_size; else qtmux->current_chunk_offset += qtmux->current_chunk_size; qtmux->current_chunk_size = 0; qtmux->current_chunk_duration = 0; } last_dts = gst_util_uint64_scale_round (pad->last_dts, atom_trak_get_timescale (pad->trak), GST_SECOND); /* fragments only deal with 1 buffer == 1 chunk (== 1 sample) */ if (pad->sample_size && !qtmux->fragment_sequence) { GstClockTime expected_timestamp; /* Constant size packets: usually raw audio (with many samples per buffer (= chunk)), but can also be fixed-packet-size codecs like ADPCM */ sample_size = pad->sample_size; if (buffer_size % sample_size != 0) goto fragmented_sample; /* note: qt raw audio storage warps it implicitly into a timewise * perfect stream, discarding buffer times. * If the difference between the current PTS and the expected one * becomes too big, we error out: there was a gap and we have no way to * represent that, causing A/V sync to be off */ expected_timestamp = gst_util_uint64_scale (pad->sample_offset, GST_SECOND, atom_trak_get_timescale (pad->trak)) + pad->first_ts; if (ABSDIFF (GST_BUFFER_DTS_OR_PTS (last_buf), expected_timestamp) > qtmux->max_raw_audio_drift) goto raw_audio_timestamp_drift; if (GST_BUFFER_DURATION (last_buf) != GST_CLOCK_TIME_NONE) { nsamples = gst_util_uint64_scale_round (GST_BUFFER_DURATION (last_buf), atom_trak_get_timescale (pad->trak), GST_SECOND); duration = GST_BUFFER_DURATION (last_buf); } else { nsamples = buffer_size / sample_size; duration = gst_util_uint64_scale_round (nsamples, GST_SECOND, atom_trak_get_timescale (pad->trak)); } /* timescale = samplerate */ scaled_duration = 1; pad->last_dts = pad->first_dts + gst_util_uint64_scale_round (pad->sample_offset + nsamples, GST_SECOND, atom_trak_get_timescale (pad->trak)); } else { nsamples = 1; sample_size = buffer_size; if (!pad->sparse && ((buf && GST_BUFFER_DTS_IS_VALID (buf)) || GST_BUFFER_DTS_IS_VALID (last_buf))) { gint64 scaled_dts; if (buf && GST_BUFFER_DTS_IS_VALID (buf)) { pad->last_dts = GST_BUFFER_DTS (buf); } else { pad->last_dts = GST_BUFFER_DTS (last_buf) + duration; } if ((gint64) (pad->last_dts) < 0) { scaled_dts = -gst_util_uint64_scale_round (-pad->last_dts, atom_trak_get_timescale (pad->trak), GST_SECOND); } else { scaled_dts = gst_util_uint64_scale_round (pad->last_dts, atom_trak_get_timescale (pad->trak), GST_SECOND); } scaled_duration = scaled_dts - last_dts; last_dts = scaled_dts; } else { /* first convert intended timestamp (in GstClockTime resolution) to * trak timescale, then derive delta; * this ensures sums of (scale)delta add up to converted timestamp, * which only deviates at most 1/scale from timestamp itself */ scaled_duration = gst_util_uint64_scale_round (pad->last_dts + duration, atom_trak_get_timescale (pad->trak), GST_SECOND) - last_dts; pad->last_dts += duration; } } gst_qt_mux_register_buffer_in_chunk (qtmux, pad, buffer_size, duration); chunk_offset = qtmux->current_chunk_offset; GST_LOG_OBJECT (qtmux, "Pad (%s) dts updated to %" GST_TIME_FORMAT, GST_PAD_NAME (pad->collect.pad), GST_TIME_ARGS (pad->last_dts)); GST_LOG_OBJECT (qtmux, "Adding %d samples to track, duration: %" G_GUINT64_FORMAT " size: %" G_GUINT32_FORMAT " chunk offset: %" G_GUINT64_FORMAT, nsamples, scaled_duration, sample_size, chunk_offset); /* might be a sync sample */ if (pad->sync && !GST_BUFFER_FLAG_IS_SET (last_buf, GST_BUFFER_FLAG_DELTA_UNIT)) { GST_LOG_OBJECT (qtmux, "Adding new sync sample entry for track of pad %s", GST_PAD_NAME (pad->collect.pad)); sync = TRUE; } if (GST_BUFFER_DTS_IS_VALID (last_buf)) { last_dts = gst_util_uint64_scale_round (GST_BUFFER_DTS (last_buf), atom_trak_get_timescale (pad->trak), GST_SECOND); pts_offset = (gint64) (gst_util_uint64_scale_round (GST_BUFFER_PTS (last_buf), atom_trak_get_timescale (pad->trak), GST_SECOND) - last_dts); } else { pts_offset = 0; last_dts = gst_util_uint64_scale_round (GST_BUFFER_PTS (last_buf), atom_trak_get_timescale (pad->trak), GST_SECOND); } GST_DEBUG ("dts: %" GST_TIME_FORMAT " pts: %" GST_TIME_FORMAT " timebase_dts: %d pts_offset: %d", GST_TIME_ARGS (GST_BUFFER_DTS (last_buf)), GST_TIME_ARGS (GST_BUFFER_PTS (last_buf)), (int) (last_dts), (int) (pts_offset)); if (GST_CLOCK_TIME_IS_VALID (duration) && (qtmux->current_chunk_duration > qtmux->longest_chunk || !GST_CLOCK_TIME_IS_VALID (qtmux->longest_chunk))) { GST_DEBUG_OBJECT (qtmux, "New longest chunk found: %" GST_TIME_FORMAT ", pad %s", GST_TIME_ARGS (qtmux->current_chunk_duration), GST_PAD_NAME (pad->collect.pad)); qtmux->longest_chunk = qtmux->current_chunk_duration; } if (qtmux->mux_mode == GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL) { const TrakBufferEntryInfo *sample_entry; guint64 block_idx = prefill_get_block_index (qtmux, pad); if (block_idx >= pad->samples->len) { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Unexpected sample %" G_GUINT64_FORMAT ", expected up to %u", block_idx, pad->samples->len)); goto bail; } /* Check if all values are as expected */ sample_entry = &g_array_index (pad->samples, TrakBufferEntryInfo, block_idx); if (chunk_offset < sample_entry->chunk_offset) { guint fill_size = sample_entry->chunk_offset - chunk_offset; GstBuffer *fill_buf; fill_buf = gst_buffer_new_allocate (NULL, fill_size, NULL); gst_buffer_memset (fill_buf, 0, 0, fill_size); ret = gst_qt_mux_send_buffer (qtmux, fill_buf, &qtmux->mdat_size, TRUE); if (ret != GST_FLOW_OK) goto bail; qtmux->current_chunk_offset = chunk_offset = sample_entry->chunk_offset; qtmux->current_chunk_size = buffer_size; qtmux->current_chunk_duration = duration; } else if (chunk_offset != sample_entry->chunk_offset) { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Unexpected chunk offset %" G_GUINT64_FORMAT ", expected up to %" G_GUINT64_FORMAT, chunk_offset, sample_entry->chunk_offset)); goto bail; } } /* now we go and register this buffer/sample all over */ ret = gst_qt_mux_register_and_push_sample (qtmux, pad, last_buf, buf == NULL, nsamples, last_dts, scaled_duration, sample_size, chunk_offset, sync, TRUE, pts_offset); pad->sample_offset += nsamples; /* if this is sparse and we have a next buffer, check if there is any gap * between them to insert an empty sample */ if (pad->sparse && buf) { if (pad->create_empty_buffer) { GstBuffer *empty_buf; gint64 empty_duration = GST_BUFFER_PTS (buf) - (GST_BUFFER_PTS (last_buf) + duration); gint64 empty_duration_scaled; guint empty_size; empty_buf = pad->create_empty_buffer (pad, empty_duration); pad->last_dts = GST_BUFFER_PTS (buf); empty_duration_scaled = gst_util_uint64_scale_round (pad->last_dts, atom_trak_get_timescale (pad->trak), GST_SECOND) - (last_dts + scaled_duration); empty_size = gst_buffer_get_size (empty_buf); gst_qt_mux_register_buffer_in_chunk (qtmux, pad, empty_size, empty_duration); ret = gst_qt_mux_register_and_push_sample (qtmux, pad, empty_buf, FALSE, 1, last_dts + scaled_duration, empty_duration_scaled, empty_size, chunk_offset, sync, TRUE, 0); } else if (pad->fourcc != FOURCC_c608 && pad->fourcc != FOURCC_c708) { /* This assert is kept here to make sure implementors of new * sparse input format decide whether there needs to be special * gap handling or not */ g_assert_not_reached (); GST_WARNING_OBJECT (qtmux, "no empty buffer creation function found for pad %s", GST_PAD_NAME (pad->collect.pad)); } } exit: return ret; /* ERRORS */ bail: { gst_buffer_unref (last_buf); return GST_FLOW_ERROR; } fragmented_sample: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Audio buffer contains fragmented sample.")); goto bail; } raw_audio_timestamp_drift: { /* TODO: Could in theory be implemented with edit lists */ GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Audio stream timestamps are drifting (got %" GST_TIME_FORMAT ", expected %" GST_TIME_FORMAT "). This is not supported yet!", GST_TIME_ARGS (GST_BUFFER_DTS_OR_PTS (last_buf)), GST_TIME_ARGS (gst_util_uint64_scale (pad->sample_offset, GST_SECOND, atom_trak_get_timescale (pad->trak)) + pad->first_ts))); goto bail; } no_pts: { GST_ELEMENT_ERROR (qtmux, STREAM, MUX, (NULL), ("Buffer has no PTS.")); goto bail; } not_negotiated: { GST_ELEMENT_ERROR (qtmux, CORE, NEGOTIATION, (NULL), ("format wasn't negotiated before buffer flow on pad %s", GST_PAD_NAME (pad->collect.pad))); if (buf) gst_buffer_unref (buf); return GST_FLOW_NOT_NEGOTIATED; } } /* * DTS running time can be negative. There is no way to represent that in * MP4 however, thus we need to offset DTS so that it starts from 0. */ static void gst_qt_pad_adjust_buffer_dts (GstQTMux * qtmux, GstQTPad * pad, GstCollectData * cdata, GstBuffer ** buf) { GstClockTime pts; gint64 dts; pts = GST_BUFFER_PTS (*buf); dts = GST_COLLECT_PADS_DTS (cdata); GST_LOG_OBJECT (qtmux, "selected pad %s with PTS %" GST_TIME_FORMAT " and DTS %" GST_STIME_FORMAT, GST_PAD_NAME (cdata->pad), GST_TIME_ARGS (pts), GST_STIME_ARGS (dts)); if (!GST_CLOCK_TIME_IS_VALID (pad->dts_adjustment)) { if (GST_CLOCK_STIME_IS_VALID (dts) && dts < 0) pad->dts_adjustment = -dts; else pad->dts_adjustment = 0; } if (pad->dts_adjustment > 0) { *buf = gst_buffer_make_writable (*buf); dts += pad->dts_adjustment; if (GST_CLOCK_TIME_IS_VALID (pts)) pts += pad->dts_adjustment; if (GST_CLOCK_STIME_IS_VALID (dts) && dts < 0) { GST_WARNING_OBJECT (pad, "Decreasing DTS."); dts = 0; } if (pts < dts) { GST_WARNING_OBJECT (pad, "DTS is bigger then PTS"); pts = dts; } GST_BUFFER_PTS (*buf) = pts; GST_BUFFER_DTS (*buf) = dts; GST_LOG_OBJECT (qtmux, "time adjusted to PTS %" GST_TIME_FORMAT " and DTS %" GST_TIME_FORMAT, GST_TIME_ARGS (pts), GST_TIME_ARGS (dts)); } } static GstQTPad * find_best_pad (GstQTMux * qtmux, GstCollectPads * pads) { GSList *walk; GstQTPad *best_pad = NULL; if (qtmux->mux_mode == GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL) { guint64 smallest_offset = G_MAXUINT64; guint64 chunk_offset = 0; for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; const TrakBufferEntryInfo *sample_entry; guint64 block_idx, current_block_idx; guint64 chunk_offset_offset = 0; GstBuffer *tmp_buf = gst_collect_pads_peek (pads, (GstCollectData *) qtpad); /* Check for EOS pads and just skip them */ if (!tmp_buf && !qtpad->last_buf && (!qtpad->raw_audio_adapter || gst_adapter_available (qtpad->raw_audio_adapter) == 0)) continue; if (tmp_buf) gst_buffer_unref (tmp_buf); /* Find the exact offset where the next sample of this track is supposed * to be written at */ block_idx = current_block_idx = prefill_get_block_index (qtmux, qtpad); sample_entry = &g_array_index (qtpad->samples, TrakBufferEntryInfo, block_idx); while (block_idx > 0) { const TrakBufferEntryInfo *tmp = &g_array_index (qtpad->samples, TrakBufferEntryInfo, block_idx - 1); if (tmp->chunk_offset != sample_entry->chunk_offset) break; chunk_offset_offset += tmp->size * tmp->nsamples; block_idx--; } /* Except for the previously selected pad being EOS we always have * qtmux->current_chunk_offset + qtmux->current_chunk_size * == * sample_entry->chunk_offset + chunk_offset_offset * for the best pad. Instead of checking that, we just return the * pad that has the smallest offset for the next to-be-written sample. */ if (sample_entry->chunk_offset + chunk_offset_offset < smallest_offset) { smallest_offset = sample_entry->chunk_offset + chunk_offset_offset; best_pad = qtpad; chunk_offset = sample_entry->chunk_offset; } } if (chunk_offset != qtmux->current_chunk_offset) { qtmux->current_pad = NULL; } return best_pad; } if (qtmux->current_pad && (qtmux->interleave_bytes != 0 || qtmux->interleave_time != 0) && (qtmux->interleave_bytes == 0 || qtmux->current_chunk_size <= qtmux->interleave_bytes) && (qtmux->interleave_time == 0 || qtmux->current_chunk_duration <= qtmux->interleave_time) && qtmux->mux_mode != GST_QT_MUX_MODE_FRAGMENTED && qtmux->mux_mode != GST_QT_MUX_MODE_FRAGMENTED_STREAMABLE) { GstBuffer *tmp_buf = gst_collect_pads_peek (pads, (GstCollectData *) qtmux->current_pad); if (tmp_buf || qtmux->current_pad->last_buf) { best_pad = qtmux->current_pad; if (tmp_buf) gst_buffer_unref (tmp_buf); GST_DEBUG_OBJECT (qtmux, "Reusing pad %s:%s", GST_DEBUG_PAD_NAME (best_pad->collect.pad)); } } else if (qtmux->collect->data->next) { /* Only switch pads if we have more than one, otherwise * we can just put everything into a single chunk and save * a few bytes of offsets */ if (qtmux->current_pad) GST_DEBUG_OBJECT (qtmux, "Switching from pad %s:%s", GST_DEBUG_PAD_NAME (qtmux->current_pad->collect.pad)); best_pad = qtmux->current_pad = NULL; } if (!best_pad) { GstClockTime best_time = GST_CLOCK_TIME_NONE; for (walk = qtmux->collect->data; walk; walk = g_slist_next (walk)) { GstCollectData *cdata = (GstCollectData *) walk->data; GstQTPad *qtpad = (GstQTPad *) cdata; GstBuffer *tmp_buf; GstClockTime timestamp; tmp_buf = gst_collect_pads_peek (pads, cdata); if (!tmp_buf) { /* This one is newly EOS now, finish it for real */ if (qtpad->last_buf) { timestamp = GST_BUFFER_DTS_OR_PTS (qtpad->last_buf); } else { continue; } } else { if (qtpad->last_buf) timestamp = GST_BUFFER_DTS_OR_PTS (qtpad->last_buf); else timestamp = GST_BUFFER_DTS_OR_PTS (tmp_buf); } if (best_pad == NULL || !GST_CLOCK_TIME_IS_VALID (best_time) || timestamp < best_time) { best_pad = qtpad; best_time = timestamp; } if (tmp_buf) gst_buffer_unref (tmp_buf); } if (best_pad) { GST_DEBUG_OBJECT (qtmux, "Choosing pad %s:%s", GST_DEBUG_PAD_NAME (best_pad->collect.pad)); } else { GST_DEBUG_OBJECT (qtmux, "No best pad: EOS"); } } return best_pad; } static GstFlowReturn gst_qt_mux_collected (GstCollectPads * pads, gpointer user_data) { GstFlowReturn ret = GST_FLOW_OK; GstQTMux *qtmux = GST_QT_MUX_CAST (user_data); GstQTPad *best_pad = NULL; if (G_UNLIKELY (qtmux->state == GST_QT_MUX_STATE_STARTED)) { if ((ret = gst_qt_mux_start_file (qtmux)) != GST_FLOW_OK) return ret; qtmux->state = GST_QT_MUX_STATE_DATA; } if (G_UNLIKELY (qtmux->state == GST_QT_MUX_STATE_EOS)) return GST_FLOW_EOS; best_pad = find_best_pad (qtmux, pads); /* clipping already converted to running time */ if (best_pad != NULL) { GstBuffer *buf = NULL; if (qtmux->mux_mode != GST_QT_MUX_MODE_ROBUST_RECORDING_PREFILL || best_pad->raw_audio_adapter == NULL || best_pad->raw_audio_adapter_pts == GST_CLOCK_TIME_NONE) buf = gst_collect_pads_pop (pads, (GstCollectData *) best_pad); g_assert (buf || best_pad->last_buf || (best_pad->raw_audio_adapter && gst_adapter_available (best_pad->raw_audio_adapter) > 0)); if (buf) gst_qt_pad_adjust_buffer_dts (qtmux, best_pad, (GstCollectData *) best_pad, &buf); ret = gst_qt_mux_add_buffer (qtmux, best_pad, buf); } else { qtmux->state = GST_QT_MUX_STATE_EOS; ret = gst_qt_mux_stop_file (qtmux); if (ret == GST_FLOW_OK) { GST_DEBUG_OBJECT (qtmux, "Pushing eos"); gst_pad_push_event (qtmux->srcpad, gst_event_new_eos ()); ret = GST_FLOW_EOS; } else { GST_WARNING_OBJECT (qtmux, "Failed to stop file: %s", gst_flow_get_name (ret)); } } return ret; } static gboolean check_field (GQuark field_id, const GValue * value, gpointer user_data) { GstStructure *structure = (GstStructure *) user_data; const GValue *other = gst_structure_id_get_value (structure, field_id); if (other == NULL) return FALSE; return gst_value_compare (value, other) == GST_VALUE_EQUAL; } static gboolean gst_qtmux_caps_is_subset_full (GstQTMux * qtmux, GstCaps * subset, GstCaps * superset) { GstStructure *sub_s = gst_caps_get_structure (subset, 0); GstStructure *sup_s = gst_caps_get_structure (superset, 0); return gst_structure_foreach (sub_s, check_field, sup_s); } /* will unref @qtmux */ static gboolean gst_qt_mux_can_renegotiate (GstQTMux * qtmux, GstPad * pad, GstCaps * caps) { GstCaps *current_caps; /* does not go well to renegotiate stream mid-way, unless * the old caps are a subset of the new one (this means upstream * added more info to the caps, as both should be 'fixed' caps) */ current_caps = gst_pad_get_current_caps (pad); g_assert (caps != NULL); if (!gst_qtmux_caps_is_subset_full (qtmux, current_caps, caps)) { gst_caps_unref (current_caps); GST_WARNING_OBJECT (qtmux, "pad %s refused renegotiation to %" GST_PTR_FORMAT, GST_PAD_NAME (pad), caps); gst_object_unref (qtmux); return FALSE; } GST_DEBUG_OBJECT (qtmux, "pad %s accepted renegotiation to %" GST_PTR_FORMAT " from %" GST_PTR_FORMAT, GST_PAD_NAME (pad), caps, current_caps); gst_object_unref (qtmux); gst_caps_unref (current_caps); return TRUE; } static gboolean gst_qt_mux_audio_sink_set_caps (GstQTPad * qtpad, GstCaps * caps) { GstPad *pad = qtpad->collect.pad; GstQTMux *qtmux = GST_QT_MUX_CAST (gst_pad_get_parent (pad)); GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); GstStructure *structure; const gchar *mimetype; gint rate, channels; const GValue *value = NULL; const GstBuffer *codec_data = NULL; GstQTMuxFormat format; AudioSampleEntry entry = { 0, }; AtomInfo *ext_atom = NULL; gint constant_size = 0; const gchar *stream_format; guint32 timescale; if (qtpad->fourcc) return gst_qt_mux_can_renegotiate (qtmux, pad, caps); GST_DEBUG_OBJECT (qtmux, "%s:%s, caps=%" GST_PTR_FORMAT, GST_DEBUG_PAD_NAME (pad), caps); qtpad->prepare_buf_func = NULL; format = qtmux_klass->format; structure = gst_caps_get_structure (caps, 0); mimetype = gst_structure_get_name (structure); /* common info */ if (!gst_structure_get_int (structure, "channels", &channels) || !gst_structure_get_int (structure, "rate", &rate)) { goto refuse_caps; } /* optional */ value = gst_structure_get_value (structure, "codec_data"); if (value != NULL) codec_data = gst_value_get_buffer (value); qtpad->is_out_of_order = FALSE; /* set common properties */ entry.sample_rate = rate; entry.channels = channels; /* default */ entry.sample_size = 16; /* this is the typical compressed case */ if (format == GST_QT_MUX_FORMAT_QT) { entry.version = 1; entry.compression_id = -2; } /* now map onto a fourcc, and some extra properties */ if (strcmp (mimetype, "audio/mpeg") == 0) { gint mpegversion = 0, mpegaudioversion = 0; gint layer = -1; gst_structure_get_int (structure, "mpegversion", &mpegversion); switch (mpegversion) { case 1: gst_structure_get_int (structure, "layer", &layer); gst_structure_get_int (structure, "mpegaudioversion", &mpegaudioversion); /* mp1/2/3 */ /* note: QuickTime player does not like mp3 either way in iso/mp4 */ if (format == GST_QT_MUX_FORMAT_QT) entry.fourcc = FOURCC__mp3; else { entry.fourcc = FOURCC_mp4a; ext_atom = build_esds_extension (qtpad->trak, ESDS_OBJECT_TYPE_MPEG1_P3, ESDS_STREAM_TYPE_AUDIO, codec_data, qtpad->avg_bitrate, qtpad->max_bitrate); } if (layer == 1) { g_warn_if_fail (format == GST_QT_MUX_FORMAT_MP4 || format == GST_QT_MUX_FORMAT_QT); entry.samples_per_packet = 384; } else if (layer == 2) { g_warn_if_fail (format == GST_QT_MUX_FORMAT_MP4 || format == GST_QT_MUX_FORMAT_QT); entry.samples_per_packet = 1152; } else { g_warn_if_fail (layer == 3); entry.samples_per_packet = (mpegaudioversion <= 1) ? 1152 : 576; } entry.bytes_per_sample = 2; break; case 4: /* check stream-format */ stream_format = gst_structure_get_string (structure, "stream-format"); if (stream_format) { if (strcmp (stream_format, "raw") != 0) { GST_WARNING_OBJECT (qtmux, "Unsupported AAC stream-format %s, " "please use 'raw'", stream_format); goto refuse_caps; } } else { GST_WARNING_OBJECT (qtmux, "No stream-format present in caps, " "assuming 'raw'"); } if (!codec_data || gst_buffer_get_size ((GstBuffer *) codec_data) < 2) { GST_WARNING_OBJECT (qtmux, "no (valid) codec_data for AAC audio"); goto refuse_caps; } else { guint8 profile; gst_buffer_extract ((GstBuffer *) codec_data, 0, &profile, 1); /* warn if not Low Complexity profile */ profile >>= 3; if (profile != 2) GST_WARNING_OBJECT (qtmux, "non-LC AAC may not run well on (Apple) QuickTime/iTunes"); } /* AAC */ entry.fourcc = FOURCC_mp4a; if (format == GST_QT_MUX_FORMAT_QT) ext_atom = build_mov_aac_extension (qtpad->trak, codec_data, qtpad->avg_bitrate, qtpad->max_bitrate); else ext_atom = build_esds_extension (qtpad->trak, ESDS_OBJECT_TYPE_MPEG4_P3, ESDS_STREAM_TYPE_AUDIO, codec_data, qtpad->avg_bitrate, qtpad->max_bitrate); break; default: break; } } else if (strcmp (mimetype, "audio/AMR") == 0) { entry.fourcc = FOURCC_samr; entry.sample_size = 16; entry.samples_per_packet = 160; entry.bytes_per_sample = 2; ext_atom = build_amr_extension (); } else if (strcmp (mimetype, "audio/AMR-WB") == 0) { entry.fourcc = FOURCC_sawb; entry.sample_size = 16; entry.samples_per_packet = 320; entry.bytes_per_sample = 2; ext_atom = build_amr_extension (); } else if (strcmp (mimetype, "audio/x-raw") == 0) { GstAudioInfo info; gst_audio_info_init (&info); if (!gst_audio_info_from_caps (&info, caps)) goto refuse_caps; /* spec has no place for a distinction in these */ if (info.finfo->width != info.finfo->depth) { GST_DEBUG_OBJECT (qtmux, "width must be same as depth!"); goto refuse_caps; } if ((info.finfo->flags & GST_AUDIO_FORMAT_FLAG_SIGNED)) { if (info.finfo->endianness == G_LITTLE_ENDIAN) entry.fourcc = FOURCC_sowt; else if (info.finfo->endianness == G_BIG_ENDIAN) entry.fourcc = FOURCC_twos; else entry.fourcc = FOURCC_sowt; /* maximum backward compatibility; only new version for > 16 bit */ if (info.finfo->depth <= 16) entry.version = 0; /* not compressed in any case */ entry.compression_id = 0; /* QT spec says: max at 16 bit even if sample size were actually larger, * however, most players (e.g. QuickTime!) seem to disagree, so ... */ entry.sample_size = info.finfo->depth; entry.bytes_per_sample = info.finfo->depth / 8; entry.samples_per_packet = 1; entry.bytes_per_packet = info.finfo->depth / 8; entry.bytes_per_frame = entry.bytes_per_packet * info.channels; } else { if (info.finfo->width == 8 && info.finfo->depth == 8) { /* fall back to old 8-bit version */ entry.fourcc = FOURCC_raw_; entry.version = 0; entry.compression_id = 0; entry.sample_size = 8; } else { GST_DEBUG_OBJECT (qtmux, "non 8-bit PCM must be signed"); goto refuse_caps; } } constant_size = (info.finfo->depth / 8) * info.channels; } else if (strcmp (mimetype, "audio/x-alaw") == 0) { entry.fourcc = FOURCC_alaw; entry.samples_per_packet = 1023; entry.bytes_per_sample = 2; } else if (strcmp (mimetype, "audio/x-mulaw") == 0) { entry.fourcc = FOURCC_ulaw; entry.samples_per_packet = 1023; entry.bytes_per_sample = 2; } else if (strcmp (mimetype, "audio/x-adpcm") == 0) { gint blocksize; if (!gst_structure_get_int (structure, "block_align", &blocksize)) { GST_DEBUG_OBJECT (qtmux, "broken caps, block_align missing"); goto refuse_caps; } /* Currently only supports WAV-style IMA ADPCM, for which the codec id is 0x11 */ entry.fourcc = MS_WAVE_FOURCC (0x11); /* 4 byte header per channel (including one sample). 2 samples per byte remaining. Simplifying gives the following (samples per block per channel) */ entry.samples_per_packet = 2 * blocksize / channels - 7; entry.bytes_per_sample = 2; entry.bytes_per_frame = blocksize; entry.bytes_per_packet = blocksize / channels; /* ADPCM has constant size packets */ constant_size = 1; /* TODO: I don't really understand why this helps, but it does! Constant * size and compression_id of -2 seem to be incompatible, and other files * in the wild use this too. */ entry.compression_id = -1; ext_atom = build_ima_adpcm_extension (channels, rate, blocksize); } else if (strcmp (mimetype, "audio/x-alac") == 0) { GstBuffer *codec_config; gint len; GstMapInfo map; entry.fourcc = FOURCC_alac; gst_buffer_map ((GstBuffer *) codec_data, &map, GST_MAP_READ); /* let's check if codec data already comes with 'alac' atom prefix */ if (!codec_data || (len = map.size) < 28) { GST_DEBUG_OBJECT (qtmux, "broken caps, codec data missing"); gst_buffer_unmap ((GstBuffer *) codec_data, &map); goto refuse_caps; } if (GST_READ_UINT32_LE (map.data + 4) == FOURCC_alac) { len -= 8; codec_config = gst_buffer_copy_region ((GstBuffer *) codec_data, GST_BUFFER_COPY_MEMORY, 8, len); } else { codec_config = gst_buffer_ref ((GstBuffer *) codec_data); } gst_buffer_unmap ((GstBuffer *) codec_data, &map); if (len != 28) { /* does not look good, but perhaps some trailing unneeded stuff */ GST_WARNING_OBJECT (qtmux, "unexpected codec-data size, possibly broken"); } if (format == GST_QT_MUX_FORMAT_QT) ext_atom = build_mov_alac_extension (codec_config); else ext_atom = build_codec_data_extension (FOURCC_alac, codec_config); /* set some more info */ gst_buffer_map (codec_config, &map, GST_MAP_READ); entry.bytes_per_sample = 2; entry.samples_per_packet = GST_READ_UINT32_BE (map.data + 4); gst_buffer_unmap (codec_config, &map); gst_buffer_unref (codec_config); } else if (strcmp (mimetype, "audio/x-ac3") == 0) { entry.fourcc = FOURCC_ac_3; /* Fixed values according to TS 102 366 but it also mentions that * they should be ignored */ entry.channels = 2; entry.sample_size = 16; /* AC-3 needs an extension atom but its data can only be obtained from * the stream itself. Abuse the prepare_buf_func so we parse a frame * and get the needed data */ qtpad->prepare_buf_func = gst_qt_mux_prepare_parse_ac3_frame; } else if (strcmp (mimetype, "audio/x-opus") == 0) { /* Based on the specification defined in: * https://www.opus-codec.org/docs/opus_in_isobmff.html */ guint8 channels, mapping_family, stream_count, coupled_count; guint16 pre_skip; gint16 output_gain; guint32 rate; guint8 channel_mapping[256]; const GValue *streamheader; const GValue *first_element; GstBuffer *header; entry.fourcc = FOURCC_opus; entry.sample_size = 16; streamheader = gst_structure_get_value (structure, "streamheader"); if (streamheader && GST_VALUE_HOLDS_ARRAY (streamheader) && gst_value_array_get_size (streamheader) != 0) { first_element = gst_value_array_get_value (streamheader, 0); header = gst_value_get_buffer (first_element); if (!gst_codec_utils_opus_parse_header (header, &rate, &channels, &mapping_family, &stream_count, &coupled_count, channel_mapping, &pre_skip, &output_gain)) { GST_ERROR_OBJECT (qtmux, "Incomplete OpusHead"); goto refuse_caps; } } else { GST_WARNING_OBJECT (qtmux, "no streamheader field in caps %" GST_PTR_FORMAT, caps); if (!gst_codec_utils_opus_parse_caps (caps, &rate, &channels, &mapping_family, &stream_count, &coupled_count, channel_mapping)) { GST_ERROR_OBJECT (qtmux, "Incomplete Opus caps"); goto refuse_caps; } pre_skip = 0; output_gain = 0; } entry.channels = channels; ext_atom = build_opus_extension (rate, channels, mapping_family, stream_count, coupled_count, channel_mapping, pre_skip, output_gain); } if (!entry.fourcc) goto refuse_caps; timescale = gst_qt_mux_pad_get_timescale (GST_QT_MUX_PAD_CAST (pad)); if (!timescale && qtmux->trak_timescale) timescale = qtmux->trak_timescale; else if (!timescale) timescale = entry.sample_rate; /* ok, set the pad info accordingly */ qtpad->fourcc = entry.fourcc; qtpad->sample_size = constant_size; qtpad->trak_ste = (SampleTableEntry *) atom_trak_set_audio_type (qtpad->trak, qtmux->context, &entry, timescale, ext_atom, constant_size); gst_object_unref (qtmux); return TRUE; /* ERRORS */ refuse_caps: { GST_WARNING_OBJECT (qtmux, "pad %s refused caps %" GST_PTR_FORMAT, GST_PAD_NAME (pad), caps); gst_object_unref (qtmux); return FALSE; } } static gboolean gst_qt_mux_video_sink_set_caps (GstQTPad * qtpad, GstCaps * caps) { GstPad *pad = qtpad->collect.pad; GstQTMux *qtmux = GST_QT_MUX_CAST (gst_pad_get_parent (pad)); GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); GstStructure *structure; const gchar *mimetype; gint width, height, depth = -1; gint framerate_num, framerate_den; guint32 rate; const GValue *value = NULL; const GstBuffer *codec_data = NULL; VisualSampleEntry entry = { 0, }; GstQTMuxFormat format; AtomInfo *ext_atom = NULL; GList *ext_atom_list = NULL; gboolean sync = FALSE; int par_num, par_den; const gchar *multiview_mode; if (qtpad->fourcc) return gst_qt_mux_can_renegotiate (qtmux, pad, caps); GST_DEBUG_OBJECT (qtmux, "%s:%s, caps=%" GST_PTR_FORMAT, GST_DEBUG_PAD_NAME (pad), caps); qtpad->prepare_buf_func = NULL; format = qtmux_klass->format; structure = gst_caps_get_structure (caps, 0); mimetype = gst_structure_get_name (structure); /* required parts */ if (!gst_structure_get_int (structure, "width", &width) || !gst_structure_get_int (structure, "height", &height)) goto refuse_caps; /* optional */ depth = -1; /* works as a default timebase */ framerate_num = 10000; framerate_den = 1; gst_structure_get_fraction (structure, "framerate", &framerate_num, &framerate_den); gst_structure_get_int (structure, "depth", &depth); value = gst_structure_get_value (structure, "codec_data"); if (value != NULL) codec_data = gst_value_get_buffer (value); par_num = 1; par_den = 1; gst_structure_get_fraction (structure, "pixel-aspect-ratio", &par_num, &par_den); qtpad->is_out_of_order = FALSE; /* bring frame numerator into a range that ensures both reasonable resolution * as well as a fair duration */ qtpad->expected_sample_duration_n = framerate_num; qtpad->expected_sample_duration_d = framerate_den; rate = gst_qt_mux_pad_get_timescale (GST_QT_MUX_PAD_CAST (pad)); if (!rate && qtmux->trak_timescale) rate = qtmux->trak_timescale; else if (!rate) rate = atom_framerate_to_timescale (framerate_num, framerate_den); GST_DEBUG_OBJECT (qtmux, "Rate of video track selected: %" G_GUINT32_FORMAT, rate); multiview_mode = gst_structure_get_string (structure, "multiview-mode"); if (multiview_mode && !qtpad->trak->mdia.minf.stbl.svmi) { GstVideoMultiviewMode mode; GstVideoMultiviewFlags flags = 0; mode = gst_video_multiview_mode_from_caps_string (multiview_mode); gst_structure_get_flagset (structure, "multiview-flags", &flags, NULL); switch (mode) { case GST_VIDEO_MULTIVIEW_MODE_SIDE_BY_SIDE: qtpad->trak->mdia.minf.stbl.svmi = atom_svmi_new (0, flags & GST_VIDEO_MULTIVIEW_FLAGS_RIGHT_VIEW_FIRST); break; case GST_VIDEO_MULTIVIEW_MODE_ROW_INTERLEAVED: qtpad->trak->mdia.minf.stbl.svmi = atom_svmi_new (1, flags & GST_VIDEO_MULTIVIEW_FLAGS_RIGHT_VIEW_FIRST); break; case GST_VIDEO_MULTIVIEW_MODE_FRAME_BY_FRAME: qtpad->trak->mdia.minf.stbl.svmi = atom_svmi_new (2, flags & GST_VIDEO_MULTIVIEW_FLAGS_RIGHT_VIEW_FIRST); break; default: GST_DEBUG_OBJECT (qtmux, "Unsupported multiview-mode %s", multiview_mode); break; } } /* set common properties */ entry.width = width; entry.height = height; entry.par_n = par_num; entry.par_d = par_den; /* should be OK according to qt and iso spec, override if really needed */ entry.color_table_id = -1; entry.frame_count = 1; entry.depth = 24; /* sync entries by default */ sync = TRUE; /* now map onto a fourcc, and some extra properties */ if (strcmp (mimetype, "video/x-raw") == 0) { const gchar *format; GstVideoFormat fmt; const GstVideoFormatInfo *vinfo; format = gst_structure_get_string (structure, "format"); fmt = gst_video_format_from_string (format); vinfo = gst_video_format_get_info (fmt); switch (fmt) { case GST_VIDEO_FORMAT_UYVY: if (depth == -1) depth = 24; entry.fourcc = FOURCC_2vuy; entry.depth = depth; sync = FALSE; break; case GST_VIDEO_FORMAT_v210: if (depth == -1) depth = 24; entry.fourcc = FOURCC_v210; entry.depth = depth; sync = FALSE; break; default: if (GST_VIDEO_FORMAT_INFO_FLAGS (vinfo) & GST_VIDEO_FORMAT_FLAG_RGB) { entry.fourcc = FOURCC_raw_; entry.depth = GST_VIDEO_FORMAT_INFO_PSTRIDE (vinfo, 0) * 8; sync = FALSE; } break; } } else if (strcmp (mimetype, "video/x-h263") == 0) { ext_atom = NULL; if (format == GST_QT_MUX_FORMAT_QT) entry.fourcc = FOURCC_h263; else entry.fourcc = FOURCC_s263; ext_atom = build_h263_extension (); if (ext_atom != NULL) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); } else if (strcmp (mimetype, "video/x-divx") == 0 || strcmp (mimetype, "video/mpeg") == 0) { gint version = 0; if (strcmp (mimetype, "video/x-divx") == 0) { gst_structure_get_int (structure, "divxversion", &version); version = version == 5 ? 1 : 0; } else { gst_structure_get_int (structure, "mpegversion", &version); version = version == 4 ? 1 : 0; } if (version) { entry.fourcc = FOURCC_mp4v; ext_atom = build_esds_extension (qtpad->trak, ESDS_OBJECT_TYPE_MPEG4_P2, ESDS_STREAM_TYPE_VISUAL, codec_data, qtpad->avg_bitrate, qtpad->max_bitrate); if (ext_atom != NULL) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); if (!codec_data) GST_WARNING_OBJECT (qtmux, "no codec_data for MPEG4 video; " "output might not play in Apple QuickTime (try global-headers?)"); } } else if (strcmp (mimetype, "video/x-h264") == 0) { if (!codec_data) { GST_WARNING_OBJECT (qtmux, "no codec_data in h264 caps"); goto refuse_caps; } entry.fourcc = FOURCC_avc1; ext_atom = build_btrt_extension (0, qtpad->avg_bitrate, qtpad->max_bitrate); if (ext_atom != NULL) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); ext_atom = build_codec_data_extension (FOURCC_avcC, codec_data); if (ext_atom != NULL) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); } else if (strcmp (mimetype, "video/x-h265") == 0) { const gchar *format; if (!codec_data) { GST_WARNING_OBJECT (qtmux, "no codec_data in h265 caps"); goto refuse_caps; } format = gst_structure_get_string (structure, "stream-format"); if (strcmp (format, "hvc1") == 0) entry.fourcc = FOURCC_hvc1; else if (strcmp (format, "hev1") == 0) entry.fourcc = FOURCC_hev1; ext_atom = build_btrt_extension (0, qtpad->avg_bitrate, qtpad->max_bitrate); if (ext_atom != NULL) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); ext_atom = build_codec_data_extension (FOURCC_hvcC, codec_data); if (ext_atom != NULL) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); } else if (strcmp (mimetype, "video/x-svq") == 0) { gint version = 0; const GstBuffer *seqh = NULL; const GValue *seqh_value; gdouble gamma = 0; gst_structure_get_int (structure, "svqversion", &version); if (version == 3) { entry.fourcc = FOURCC_SVQ3; entry.version = 3; entry.depth = 32; seqh_value = gst_structure_get_value (structure, "seqh"); if (seqh_value) { seqh = gst_value_get_buffer (seqh_value); ext_atom = build_SMI_atom (seqh); if (ext_atom) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); } /* we need to add the gamma anyway because quicktime might crash * when it doesn't find it */ if (!gst_structure_get_double (structure, "applied-gamma", &gamma)) { /* it seems that using 0 here makes it ignored */ gamma = 0.0; } ext_atom = build_gama_atom (gamma); if (ext_atom) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); } else { GST_WARNING_OBJECT (qtmux, "SVQ version %d not supported. Please file " "a bug at http://bugzilla.gnome.org", version); } } else if (strcmp (mimetype, "video/x-dv") == 0) { gint version = 0; gboolean pal = TRUE; sync = FALSE; if (framerate_num != 25 || framerate_den != 1) pal = FALSE; gst_structure_get_int (structure, "dvversion", &version); /* fall back to typical one */ if (!version) version = 25; switch (version) { case 25: if (pal) entry.fourcc = FOURCC_dvcp; else entry.fourcc = FOURCC_dvc_; break; case 50: if (pal) entry.fourcc = FOURCC_dv5p; else entry.fourcc = FOURCC_dv5n; break; default: GST_WARNING_OBJECT (qtmux, "unrecognized dv version"); break; } } else if (strcmp (mimetype, "image/jpeg") == 0) { entry.fourcc = FOURCC_jpeg; sync = FALSE; } else if (strcmp (mimetype, "image/png") == 0) { entry.fourcc = FOURCC_png; sync = FALSE; } else if (strcmp (mimetype, "image/x-j2c") == 0 || strcmp (mimetype, "image/x-jpc") == 0) { const gchar *colorspace; const GValue *cmap_array; const GValue *cdef_array; gint ncomp = 0; if (strcmp (mimetype, "image/x-jpc") == 0) { qtpad->prepare_buf_func = gst_qt_mux_prepare_jpc_buffer; } gst_structure_get_int (structure, "num-components", &ncomp); cmap_array = gst_structure_get_value (structure, "component-map"); cdef_array = gst_structure_get_value (structure, "channel-definitions"); ext_atom = NULL; entry.fourcc = FOURCC_mjp2; sync = FALSE; colorspace = gst_structure_get_string (structure, "colorspace"); if (colorspace && (ext_atom = build_jp2h_extension (width, height, colorspace, ncomp, cmap_array, cdef_array)) != NULL) { ext_atom_list = g_list_append (ext_atom_list, ext_atom); ext_atom = build_jp2x_extension (codec_data); if (ext_atom) ext_atom_list = g_list_append (ext_atom_list, ext_atom); } else { GST_DEBUG_OBJECT (qtmux, "missing or invalid fourcc in jp2 caps"); goto refuse_caps; } } else if (strcmp (mimetype, "video/x-vp8") == 0) { entry.fourcc = FOURCC_vp08; } else if (strcmp (mimetype, "video/x-vp9") == 0) { entry.fourcc = FOURCC_vp09; } else if (strcmp (mimetype, "video/x-dirac") == 0) { entry.fourcc = FOURCC_drac; } else if (strcmp (mimetype, "video/x-qt-part") == 0) { guint32 fourcc = 0; gst_structure_get_uint (structure, "format", &fourcc); entry.fourcc = fourcc; } else if (strcmp (mimetype, "video/x-mp4-part") == 0) { guint32 fourcc = 0; gst_structure_get_uint (structure, "format", &fourcc); entry.fourcc = fourcc; } else if (strcmp (mimetype, "video/x-prores") == 0) { const gchar *variant; variant = gst_structure_get_string (structure, "variant"); if (!variant || !g_strcmp0 (variant, "standard")) entry.fourcc = FOURCC_apcn; else if (!g_strcmp0 (variant, "lt")) entry.fourcc = FOURCC_apcs; else if (!g_strcmp0 (variant, "hq")) entry.fourcc = FOURCC_apch; else if (!g_strcmp0 (variant, "proxy")) entry.fourcc = FOURCC_apco; else if (!g_strcmp0 (variant, "4444")) entry.fourcc = FOURCC_ap4h; else if (!g_strcmp0 (variant, "4444xq")) entry.fourcc = FOURCC_ap4x; sync = FALSE; if (!qtmux->interleave_time_set) qtmux->interleave_time = 500 * GST_MSECOND; if (!qtmux->interleave_bytes_set) qtmux->interleave_bytes = width > 720 ? 4 * 1024 * 1024 : 2 * 1024 * 1024; } else if (strcmp (mimetype, "video/x-cineform") == 0) { entry.fourcc = FOURCC_cfhd; sync = FALSE; } else if (strcmp (mimetype, "video/x-av1") == 0) { gint presentation_delay; guint8 presentation_delay_byte = 0; GstBuffer *av1_codec_data; if (gst_structure_get_int (structure, "presentation-delay", &presentation_delay)) { presentation_delay_byte = 1 << 5; presentation_delay_byte |= MAX (0xF, presentation_delay & 0xF); } av1_codec_data = gst_buffer_new_allocate (NULL, 5, NULL); /* Fill version and 3 bytes of flags to 0 */ gst_buffer_memset (av1_codec_data, 0, 0, 4); gst_buffer_fill (av1_codec_data, 4, &presentation_delay_byte, 1); if (codec_data) av1_codec_data = gst_buffer_append (av1_codec_data, gst_buffer_ref ((GstBuffer *) codec_data)); entry.fourcc = FOURCC_av01; ext_atom = build_btrt_extension (0, qtpad->avg_bitrate, qtpad->max_bitrate); if (ext_atom != NULL) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); ext_atom = build_codec_data_extension (FOURCC_av1C, av1_codec_data); if (ext_atom != NULL) ext_atom_list = g_list_prepend (ext_atom_list, ext_atom); gst_buffer_unref (av1_codec_data); } if (!entry.fourcc) goto refuse_caps; if (qtmux_klass->format == GST_QT_MUX_FORMAT_QT || qtmux_klass->format == GST_QT_MUX_FORMAT_MP4) { const gchar *s; GstVideoColorimetry colorimetry; s = gst_structure_get_string (structure, "colorimetry"); if (s && gst_video_colorimetry_from_string (&colorimetry, s)) { ext_atom = build_colr_extension (&colorimetry, qtmux_klass->format == GST_QT_MUX_FORMAT_MP4); if (ext_atom) ext_atom_list = g_list_append (ext_atom_list, ext_atom); } } if (qtmux_klass->format == GST_QT_MUX_FORMAT_QT || strcmp (mimetype, "image/x-j2c") == 0 || strcmp (mimetype, "image/x-jpc") == 0) { const gchar *s; GstVideoInterlaceMode interlace_mode; GstVideoFieldOrder field_order; gint fields = -1; if (strcmp (mimetype, "image/x-j2c") == 0 || strcmp (mimetype, "image/x-jpc") == 0) { fields = 1; gst_structure_get_int (structure, "fields", &fields); } s = gst_structure_get_string (structure, "interlace-mode"); if (s) interlace_mode = gst_video_interlace_mode_from_string (s); else interlace_mode = (fields <= 1) ? GST_VIDEO_INTERLACE_MODE_PROGRESSIVE : GST_VIDEO_INTERLACE_MODE_MIXED; field_order = GST_VIDEO_FIELD_ORDER_UNKNOWN; if (interlace_mode == GST_VIDEO_INTERLACE_MODE_INTERLEAVED) { s = gst_structure_get_string (structure, "field-order"); if (s) field_order = gst_video_field_order_from_string (s); } ext_atom = build_fiel_extension (interlace_mode, field_order); if (ext_atom) ext_atom_list = g_list_append (ext_atom_list, ext_atom); } if (qtmux_klass->format == GST_QT_MUX_FORMAT_QT && width > 640 && width <= 1052 && height >= 480 && height <= 576) { /* The 'clap' extension is also defined for MP4 but inventing values in * general seems a bit tricky for this one. We only write it for * SD resolution in MOV, where it is a requirement. * The same goes for the 'tapt' extension, just that it is not defined for * MP4 and only for MOV */ gint dar_num, dar_den; gint clef_width, clef_height, prof_width; gint clap_width_n, clap_width_d, clap_height; gint cdiv; double approx_dar; /* First, guess display aspect ratio based on pixel aspect ratio, * width and height. We assume that display aspect ratio is either * 4:3 or 16:9 */ approx_dar = (gdouble) (width * par_num) / (height * par_den); if (approx_dar > 11.0 / 9 && approx_dar < 14.0 / 9) { dar_num = 4; dar_den = 3; } else if (approx_dar > 15.0 / 9 && approx_dar < 18.0 / 9) { dar_num = 16; dar_den = 9; } else { dar_num = width * par_num; dar_den = height * par_den; cdiv = gst_util_greatest_common_divisor (dar_num, dar_den); dar_num /= cdiv; dar_den /= cdiv; } /* Then, calculate clean-aperture values (clap and clef) * using the guessed DAR. */ clef_height = clap_height = (height == 486 ? 480 : height); clef_width = gst_util_uint64_scale (clef_height, dar_num * G_GUINT64_CONSTANT (65536), dar_den); prof_width = gst_util_uint64_scale (width, par_num * G_GUINT64_CONSTANT (65536), par_den); clap_width_n = clap_height * dar_num * par_den; clap_width_d = dar_den * par_num; cdiv = gst_util_greatest_common_divisor (clap_width_n, clap_width_d); clap_width_n /= cdiv; clap_width_d /= cdiv; ext_atom = build_tapt_extension (clef_width, clef_height << 16, prof_width, height << 16, width << 16, height << 16); qtpad->trak->tapt = ext_atom; ext_atom = build_clap_extension (clap_width_n, clap_width_d, clap_height, 1, 0, 1, 0, 1); if (ext_atom) ext_atom_list = g_list_append (ext_atom_list, ext_atom); } /* ok, set the pad info accordingly */ qtpad->fourcc = entry.fourcc; qtpad->sync = sync; qtpad->trak_ste = (SampleTableEntry *) atom_trak_set_video_type (qtpad->trak, qtmux->context, &entry, rate, ext_atom_list); if (strcmp (mimetype, "video/x-prores") == 0) { SampleTableEntryMP4V *mp4v = (SampleTableEntryMP4V *) qtpad->trak_ste; const gchar *compressor = NULL; mp4v->spatial_quality = 0x3FF; mp4v->temporal_quality = 0; mp4v->vendor = FOURCC_appl; mp4v->horizontal_resolution = 72 << 16; mp4v->vertical_resolution = 72 << 16; mp4v->depth = (entry.fourcc == FOURCC_ap4h || entry.fourcc == FOURCC_ap4x) ? 32 : 24; /* Set compressor name, required by some software */ switch (entry.fourcc) { case FOURCC_apcn: compressor = "Apple ProRes 422"; break; case FOURCC_apcs: compressor = "Apple ProRes 422 LT"; break; case FOURCC_apch: compressor = "Apple ProRes 422 HQ"; break; case FOURCC_apco: compressor = "Apple ProRes 422 Proxy"; break; case FOURCC_ap4h: compressor = "Apple ProRes 4444"; break; case FOURCC_ap4x: compressor = "Apple ProRes 4444 XQ"; break; } if (compressor) { strcpy ((gchar *) mp4v->compressor + 1, compressor); mp4v->compressor[0] = strlen (compressor); } } gst_object_unref (qtmux); return TRUE; /* ERRORS */ refuse_caps: { GST_WARNING_OBJECT (qtmux, "pad %s refused caps %" GST_PTR_FORMAT, GST_PAD_NAME (pad), caps); gst_object_unref (qtmux); return FALSE; } } static gboolean gst_qt_mux_subtitle_sink_set_caps (GstQTPad * qtpad, GstCaps * caps) { GstPad *pad = qtpad->collect.pad; GstQTMux *qtmux = GST_QT_MUX_CAST (gst_pad_get_parent (pad)); GstStructure *structure; SubtitleSampleEntry entry = { 0, }; if (qtpad->fourcc) return gst_qt_mux_can_renegotiate (qtmux, pad, caps); GST_DEBUG_OBJECT (qtmux, "%s:%s, caps=%" GST_PTR_FORMAT, GST_DEBUG_PAD_NAME (pad), caps); /* subtitles default */ subtitle_sample_entry_init (&entry); qtpad->is_out_of_order = FALSE; qtpad->sync = FALSE; qtpad->sparse = TRUE; qtpad->prepare_buf_func = NULL; structure = gst_caps_get_structure (caps, 0); if (gst_structure_has_name (structure, "text/x-raw")) { const gchar *format = gst_structure_get_string (structure, "format"); if (format && strcmp (format, "utf8") == 0) { entry.fourcc = FOURCC_tx3g; qtpad->prepare_buf_func = gst_qt_mux_prepare_tx3g_buffer; qtpad->create_empty_buffer = gst_qt_mux_create_empty_tx3g_buffer; } } if (!entry.fourcc) goto refuse_caps; qtpad->fourcc = entry.fourcc; qtpad->trak_ste = (SampleTableEntry *) atom_trak_set_subtitle_type (qtpad->trak, qtmux->context, &entry); gst_object_unref (qtmux); return TRUE; /* ERRORS */ refuse_caps: { GST_WARNING_OBJECT (qtmux, "pad %s refused caps %" GST_PTR_FORMAT, GST_PAD_NAME (pad), caps); gst_object_unref (qtmux); return FALSE; } } static gboolean gst_qt_mux_caption_sink_set_caps (GstQTPad * qtpad, GstCaps * caps) { GstPad *pad = qtpad->collect.pad; GstQTMux *qtmux = GST_QT_MUX_CAST (gst_pad_get_parent (pad)); GstStructure *structure; guint32 fourcc_entry; guint32 timescale; if (qtpad->fourcc) return gst_qt_mux_can_renegotiate (qtmux, pad, caps); GST_DEBUG_OBJECT (qtmux, "%s:%s, caps=%" GST_PTR_FORMAT, GST_DEBUG_PAD_NAME (pad), caps); /* captions default */ qtpad->is_out_of_order = FALSE; qtpad->sync = FALSE; qtpad->sparse = TRUE; /* Closed caption data are within atoms */ qtpad->prepare_buf_func = gst_qt_mux_prepare_caption_buffer; structure = gst_caps_get_structure (caps, 0); /* We know we only handle 608,format=s334-1a and 708,format=cdp */ if (gst_structure_has_name (structure, "closedcaption/x-cea-608")) { fourcc_entry = FOURCC_c608; } else if (gst_structure_has_name (structure, "closedcaption/x-cea-708")) { fourcc_entry = FOURCC_c708; } else goto refuse_caps; /* We set the real timescale later to the one from the video track when * writing the headers */ timescale = gst_qt_mux_pad_get_timescale (GST_QT_MUX_PAD_CAST (pad)); if (!timescale && qtmux->trak_timescale) timescale = qtmux->trak_timescale; else if (!timescale) timescale = 30000; qtpad->fourcc = fourcc_entry; qtpad->trak_ste = (SampleTableEntry *) atom_trak_set_caption_type (qtpad->trak, qtmux->context, timescale, fourcc_entry); /* Initialize caption track language code to 0 unless something else is * specified. Without this, Final Cut considers it "non-standard" */ qtpad->trak->mdia.mdhd.language_code = 0; gst_object_unref (qtmux); return TRUE; /* ERRORS */ refuse_caps: { GST_WARNING_OBJECT (qtmux, "pad %s refused caps %" GST_PTR_FORMAT, GST_PAD_NAME (pad), caps); gst_object_unref (qtmux); return FALSE; } } static gboolean gst_qt_mux_sink_event (GstCollectPads * pads, GstCollectData * data, GstEvent * event, gpointer user_data) { GstQTMux *qtmux; guint32 avg_bitrate = 0, max_bitrate = 0; GstPad *pad = data->pad; gboolean ret = TRUE; qtmux = GST_QT_MUX_CAST (user_data); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_CAPS: { GstCaps *caps; GstQTPad *collect_pad; gst_event_parse_caps (event, &caps); /* find stream data */ collect_pad = (GstQTPad *) gst_pad_get_element_private (pad); g_assert (collect_pad); g_assert (collect_pad->set_caps); ret = collect_pad->set_caps (collect_pad, caps); gst_event_unref (event); event = NULL; break; } case GST_EVENT_TAG:{ GstTagList *list; GstTagSetter *setter = GST_TAG_SETTER (qtmux); GstTagMergeMode mode; gchar *code; GstQTPad *collect_pad; GST_OBJECT_LOCK (qtmux); mode = gst_tag_setter_get_tag_merge_mode (setter); collect_pad = (GstQTPad *) gst_pad_get_element_private (pad); gst_event_parse_tag (event, &list); GST_DEBUG_OBJECT (qtmux, "received tag event on pad %s:%s : %" GST_PTR_FORMAT, GST_DEBUG_PAD_NAME (pad), list); if (gst_tag_list_get_scope (list) == GST_TAG_SCOPE_GLOBAL) { gst_tag_setter_merge_tags (setter, list, mode); qtmux->tags_changed = TRUE; } else { if (!collect_pad->tags) collect_pad->tags = gst_tag_list_new_empty (); gst_tag_list_insert (collect_pad->tags, list, mode); collect_pad->tags_changed = TRUE; } GST_OBJECT_UNLOCK (qtmux); if (gst_tag_list_get_uint (list, GST_TAG_BITRATE, &avg_bitrate) | gst_tag_list_get_uint (list, GST_TAG_MAXIMUM_BITRATE, &max_bitrate)) { GstQTPad *qtpad = gst_pad_get_element_private (pad); g_assert (qtpad); if (avg_bitrate > 0 && avg_bitrate < G_MAXUINT32) qtpad->avg_bitrate = avg_bitrate; if (max_bitrate > 0 && max_bitrate < G_MAXUINT32) qtpad->max_bitrate = max_bitrate; } if (gst_tag_list_get_string (list, GST_TAG_LANGUAGE_CODE, &code)) { const char *iso_code = gst_tag_get_language_code_iso_639_2T (code); if (iso_code) { GstQTPad *qtpad = gst_pad_get_element_private (pad); g_assert (qtpad); if (qtpad->trak) { /* https://developer.apple.com/library/mac/#documentation/QuickTime/QTFF/QTFFChap4/qtff4.html */ qtpad->trak->mdia.mdhd.language_code = language_code (iso_code); } } g_free (code); } gst_event_unref (event); event = NULL; ret = TRUE; break; } default: break; } if (event != NULL) return gst_collect_pads_event_default (pads, data, event, FALSE); return ret; } static void gst_qt_mux_release_pad (GstElement * element, GstPad * pad) { GstQTMux *mux = GST_QT_MUX_CAST (element); GSList *walk; GST_DEBUG_OBJECT (element, "Releasing %s:%s", GST_DEBUG_PAD_NAME (pad)); for (walk = mux->sinkpads; walk; walk = g_slist_next (walk)) { GstQTPad *qtpad = (GstQTPad *) walk->data; GST_DEBUG ("Checking %s:%s", GST_DEBUG_PAD_NAME (qtpad->collect.pad)); if (qtpad->collect.pad == pad) { /* this is it, remove */ mux->sinkpads = g_slist_delete_link (mux->sinkpads, walk); gst_element_remove_pad (element, pad); break; } } if (mux->current_pad && mux->current_pad->collect.pad == pad) { mux->current_pad = NULL; mux->current_chunk_size = 0; mux->current_chunk_duration = 0; } gst_collect_pads_remove_pad (mux->collect, pad); if (mux->sinkpads == NULL) { /* No more outstanding request pads, reset our counters */ mux->video_pads = 0; mux->audio_pads = 0; mux->subtitle_pads = 0; } } static GstPad * gst_qt_mux_request_new_pad (GstElement * element, GstPadTemplate * templ, const gchar * req_name, const GstCaps * caps) { GstElementClass *klass = GST_ELEMENT_GET_CLASS (element); GstQTMux *qtmux = GST_QT_MUX_CAST (element); GstQTPad *collect_pad; GstPad *newpad; GstQTPadSetCapsFunc setcaps_func; gchar *name; gint pad_id; gboolean lock = TRUE; if (templ->direction != GST_PAD_SINK) goto wrong_direction; if (qtmux->state > GST_QT_MUX_STATE_STARTED) goto too_late; if (templ == gst_element_class_get_pad_template (klass, "audio_%u")) { setcaps_func = gst_qt_mux_audio_sink_set_caps; if (req_name != NULL && sscanf (req_name, "audio_%u", &pad_id) == 1) { name = g_strdup (req_name); } else { name = g_strdup_printf ("audio_%u", qtmux->audio_pads++); } } else if (templ == gst_element_class_get_pad_template (klass, "video_%u")) { setcaps_func = gst_qt_mux_video_sink_set_caps; if (req_name != NULL && sscanf (req_name, "video_%u", &pad_id) == 1) { name = g_strdup (req_name); } else { name = g_strdup_printf ("video_%u", qtmux->video_pads++); } } else if (templ == gst_element_class_get_pad_template (klass, "subtitle_%u")) { setcaps_func = gst_qt_mux_subtitle_sink_set_caps; if (req_name != NULL && sscanf (req_name, "subtitle_%u", &pad_id) == 1) { name = g_strdup (req_name); } else { name = g_strdup_printf ("subtitle_%u", qtmux->subtitle_pads++); } lock = FALSE; } else if (templ == gst_element_class_get_pad_template (klass, "caption_%u")) { setcaps_func = gst_qt_mux_caption_sink_set_caps; if (req_name != NULL && sscanf (req_name, "caption_%u", &pad_id) == 1) { name = g_strdup (req_name); } else { name = g_strdup_printf ("caption_%u", qtmux->caption_pads++); } lock = FALSE; } else goto wrong_template; GST_DEBUG_OBJECT (qtmux, "Requested pad: %s", name); /* create pad and add to collections */ newpad = g_object_new (GST_TYPE_QT_MUX_PAD, "name", name, "direction", templ->direction, "template", templ, NULL); g_free (name); collect_pad = (GstQTPad *) gst_collect_pads_add_pad (qtmux->collect, newpad, sizeof (GstQTPad), (GstCollectDataDestroyNotify) (gst_qt_mux_pad_reset), lock); /* set up pad */ gst_qt_mux_pad_reset (collect_pad); collect_pad->trak = atom_trak_new (qtmux->context); atom_moov_add_trak (qtmux->moov, collect_pad->trak); qtmux->sinkpads = g_slist_append (qtmux->sinkpads, collect_pad); /* set up pad functions */ collect_pad->set_caps = setcaps_func; gst_pad_set_active (newpad, TRUE); gst_element_add_pad (element, newpad); return newpad; /* ERRORS */ wrong_direction: { GST_WARNING_OBJECT (qtmux, "Request pad that is not a SINK pad."); return NULL; } too_late: { GST_WARNING_OBJECT (qtmux, "Not providing request pad after stream start."); return NULL; } wrong_template: { GST_WARNING_OBJECT (qtmux, "This is not our template!"); return NULL; } } static void gst_qt_mux_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstQTMux *qtmux = GST_QT_MUX_CAST (object); GST_OBJECT_LOCK (qtmux); switch (prop_id) { case PROP_MOVIE_TIMESCALE: g_value_set_uint (value, qtmux->timescale); break; case PROP_TRAK_TIMESCALE: g_value_set_uint (value, qtmux->trak_timescale); break; case PROP_DO_CTTS: g_value_set_boolean (value, qtmux->guess_pts); break; #ifndef GST_REMOVE_DEPRECATED case PROP_DTS_METHOD: g_value_set_enum (value, qtmux->dts_method); break; #endif case PROP_FAST_START: g_value_set_boolean (value, qtmux->fast_start); break; case PROP_FAST_START_TEMP_FILE: g_value_set_string (value, qtmux->fast_start_file_path); break; case PROP_MOOV_RECOV_FILE: g_value_set_string (value, qtmux->moov_recov_file_path); break; case PROP_FRAGMENT_DURATION: g_value_set_uint (value, qtmux->fragment_duration); break; case PROP_STREAMABLE: g_value_set_boolean (value, qtmux->streamable); break; case PROP_RESERVED_MAX_DURATION: g_value_set_uint64 (value, qtmux->reserved_max_duration); break; case PROP_RESERVED_DURATION_REMAINING: if (qtmux->reserved_duration_remaining == GST_CLOCK_TIME_NONE) g_value_set_uint64 (value, qtmux->reserved_max_duration); else { GstClockTime remaining = qtmux->reserved_duration_remaining; /* Report the remaining space as the calculated remaining, minus * however much we've muxed since the last update */ if (remaining > qtmux->muxed_since_last_update) remaining -= qtmux->muxed_since_last_update; else remaining = 0; GST_LOG_OBJECT (qtmux, "reserved duration remaining - reporting %" G_GUINT64_FORMAT "(%" G_GUINT64_FORMAT " - %" G_GUINT64_FORMAT, remaining, qtmux->reserved_duration_remaining, qtmux->muxed_since_last_update); g_value_set_uint64 (value, remaining); } break; case PROP_RESERVED_MOOV_UPDATE_PERIOD: g_value_set_uint64 (value, qtmux->reserved_moov_update_period); break; case PROP_RESERVED_BYTES_PER_SEC: g_value_set_uint (value, qtmux->reserved_bytes_per_sec_per_trak); break; case PROP_RESERVED_PREFILL: g_value_set_boolean (value, qtmux->reserved_prefill); break; case PROP_INTERLEAVE_BYTES: g_value_set_uint64 (value, qtmux->interleave_bytes); break; case PROP_INTERLEAVE_TIME: g_value_set_uint64 (value, qtmux->interleave_time); break; case PROP_MAX_RAW_AUDIO_DRIFT: g_value_set_uint64 (value, qtmux->max_raw_audio_drift); break; case PROP_START_GAP_THRESHOLD: g_value_set_uint64 (value, qtmux->start_gap_threshold); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } GST_OBJECT_UNLOCK (qtmux); } static void gst_qt_mux_generate_fast_start_file_path (GstQTMux * qtmux) { gchar *tmp; g_free (qtmux->fast_start_file_path); qtmux->fast_start_file_path = NULL; tmp = g_strdup_printf ("%s%d", "qtmux", g_random_int ()); qtmux->fast_start_file_path = g_build_filename (g_get_tmp_dir (), tmp, NULL); g_free (tmp); } static void gst_qt_mux_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstQTMux *qtmux = GST_QT_MUX_CAST (object); GST_OBJECT_LOCK (qtmux); switch (prop_id) { case PROP_MOVIE_TIMESCALE: qtmux->timescale = g_value_get_uint (value); break; case PROP_TRAK_TIMESCALE: qtmux->trak_timescale = g_value_get_uint (value); break; case PROP_DO_CTTS: qtmux->guess_pts = g_value_get_boolean (value); break; #ifndef GST_REMOVE_DEPRECATED case PROP_DTS_METHOD: qtmux->dts_method = g_value_get_enum (value); break; #endif case PROP_FAST_START: qtmux->fast_start = g_value_get_boolean (value); break; case PROP_FAST_START_TEMP_FILE: g_free (qtmux->fast_start_file_path); qtmux->fast_start_file_path = g_value_dup_string (value); /* NULL means to generate a random one */ if (!qtmux->fast_start_file_path) { gst_qt_mux_generate_fast_start_file_path (qtmux); } break; case PROP_MOOV_RECOV_FILE: g_free (qtmux->moov_recov_file_path); qtmux->moov_recov_file_path = g_value_dup_string (value); break; case PROP_FRAGMENT_DURATION: qtmux->fragment_duration = g_value_get_uint (value); break; case PROP_STREAMABLE:{ GstQTMuxClass *qtmux_klass = (GstQTMuxClass *) (G_OBJECT_GET_CLASS (qtmux)); if (qtmux_klass->format == GST_QT_MUX_FORMAT_ISML) { qtmux->streamable = g_value_get_boolean (value); } break; } case PROP_RESERVED_MAX_DURATION: qtmux->reserved_max_duration = g_value_get_uint64 (value); break; case PROP_RESERVED_MOOV_UPDATE_PERIOD: qtmux->reserved_moov_update_period = g_value_get_uint64 (value); break; case PROP_RESERVED_BYTES_PER_SEC: qtmux->reserved_bytes_per_sec_per_trak = g_value_get_uint (value); break; case PROP_RESERVED_PREFILL: qtmux->reserved_prefill = g_value_get_boolean (value); break; case PROP_INTERLEAVE_BYTES: qtmux->interleave_bytes = g_value_get_uint64 (value); qtmux->interleave_bytes_set = TRUE; break; case PROP_INTERLEAVE_TIME: qtmux->interleave_time = g_value_get_uint64 (value); qtmux->interleave_time_set = TRUE; break; case PROP_MAX_RAW_AUDIO_DRIFT: qtmux->max_raw_audio_drift = g_value_get_uint64 (value); break; case PROP_START_GAP_THRESHOLD: qtmux->start_gap_threshold = g_value_get_uint64 (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } GST_OBJECT_UNLOCK (qtmux); } static GstStateChangeReturn gst_qt_mux_change_state (GstElement * element, GstStateChange transition) { GstStateChangeReturn ret; GstQTMux *qtmux = GST_QT_MUX_CAST (element); switch (transition) { case GST_STATE_CHANGE_NULL_TO_READY: break; case GST_STATE_CHANGE_READY_TO_PAUSED: gst_collect_pads_start (qtmux->collect); qtmux->state = GST_QT_MUX_STATE_STARTED; break; case GST_STATE_CHANGE_PAUSED_TO_PLAYING: break; case GST_STATE_CHANGE_PAUSED_TO_READY: gst_collect_pads_stop (qtmux->collect); break; default: break; } ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); switch (transition) { case GST_STATE_CHANGE_PLAYING_TO_PAUSED: break; case GST_STATE_CHANGE_PAUSED_TO_READY: gst_qt_mux_reset (qtmux, TRUE); break; case GST_STATE_CHANGE_READY_TO_NULL: break; default: break; } return ret; } gboolean gst_qt_mux_register (GstPlugin * plugin) { GTypeInfo typeinfo = { sizeof (GstQTMuxClass), (GBaseInitFunc) gst_qt_mux_base_init, NULL, (GClassInitFunc) gst_qt_mux_class_init, NULL, NULL, sizeof (GstQTMux), 0, (GInstanceInitFunc) gst_qt_mux_init, }; static const GInterfaceInfo tag_setter_info = { NULL, NULL, NULL }; static const GInterfaceInfo tag_xmp_writer_info = { NULL, NULL, NULL }; static const GInterfaceInfo preset_info = { NULL, NULL, NULL }; GType type; GstQTMuxFormat format; GstQTMuxClassParams *params; guint i = 0; GST_DEBUG_CATEGORY_INIT (gst_qt_mux_debug, "qtmux", 0, "QT Muxer"); GST_LOG ("Registering muxers"); while (TRUE) { GstQTMuxFormatProp *prop; GstCaps *subtitle_caps, *caption_caps; prop = &gst_qt_mux_format_list[i]; format = prop->format; if (format == GST_QT_MUX_FORMAT_NONE) break; /* create a cache for these properties */ params = g_new0 (GstQTMuxClassParams, 1); params->prop = prop; params->src_caps = gst_static_caps_get (&prop->src_caps); params->video_sink_caps = gst_static_caps_get (&prop->video_sink_caps); params->audio_sink_caps = gst_static_caps_get (&prop->audio_sink_caps); subtitle_caps = gst_static_caps_get (&prop->subtitle_sink_caps); if (!gst_caps_is_equal (subtitle_caps, GST_CAPS_NONE)) { params->subtitle_sink_caps = subtitle_caps; } else { gst_caps_unref (subtitle_caps); } caption_caps = gst_static_caps_get (&prop->caption_sink_caps); if (!gst_caps_is_equal (caption_caps, GST_CAPS_NONE)) { params->caption_sink_caps = caption_caps; } else { gst_caps_unref (caption_caps); } /* create the type now */ type = g_type_register_static (GST_TYPE_ELEMENT, prop->type_name, &typeinfo, 0); g_type_set_qdata (type, GST_QT_MUX_PARAMS_QDATA, (gpointer) params); g_type_add_interface_static (type, GST_TYPE_TAG_SETTER, &tag_setter_info); g_type_add_interface_static (type, GST_TYPE_TAG_XMP_WRITER, &tag_xmp_writer_info); g_type_add_interface_static (type, GST_TYPE_PRESET, &preset_info); if (!gst_element_register (plugin, prop->name, prop->rank, type)) return FALSE; i++; } GST_LOG ("Finished registering muxers"); /* FIXME: ideally classification tag should be added and registered in gstreamer core gsttaglist */ GST_LOG ("Registering tags"); gst_tag_register (GST_TAG_3GP_CLASSIFICATION, GST_TAG_FLAG_META, G_TYPE_STRING, GST_TAG_3GP_CLASSIFICATION, "content classification", gst_tag_merge_use_first); GST_LOG ("Finished registering tags"); return TRUE; }