/* GStreamer ASF/WMV/WMA demuxer * Copyright (C) <1999> Erik Walthinsen * Copyright (C) <2006> Tim-Philipp Müller * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include #include "gstasfdemux.h" #include "asfheaders.h" static GstStaticPadTemplate gst_asf_demux_sink_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS ("video/x-ms-asf") ); static GstStaticPadTemplate audio_src_template = GST_STATIC_PAD_TEMPLATE ("audio_%02d", GST_PAD_SRC, GST_PAD_SOMETIMES, GST_STATIC_CAPS_ANY); static GstStaticPadTemplate video_src_template = GST_STATIC_PAD_TEMPLATE ("video_%02d", GST_PAD_SRC, GST_PAD_SOMETIMES, GST_STATIC_CAPS_ANY); /* abuse this GstFlowReturn enum for internal usage */ #define ASF_FLOW_NEED_MORE_DATA 99 #define gst_asf_get_flow_name(flow) \ (flow == ASF_FLOW_NEED_MORE_DATA) ? \ "need-more-data" : gst_flow_get_name (flow) GST_DEBUG_CATEGORY_STATIC (asf_debug); #define GST_CAT_DEFAULT asf_debug static GstStateChangeReturn gst_asf_demux_change_state (GstElement * element, GstStateChange transition); static gboolean gst_asf_demux_element_send_event (GstElement * element, GstEvent * event); static gboolean gst_asf_demux_send_event_unlocked (GstASFDemux * demux, GstEvent * event); static gboolean gst_asf_demux_handle_src_query (GstPad * pad, GstQuery * query); static const GstQueryType *gst_asf_demux_get_src_query_types (GstPad * pad); static GstFlowReturn gst_asf_demux_parse_data (GstASFDemux * demux); static GstFlowReturn gst_asf_demux_chain (GstPad * pad, GstBuffer * buf); static gboolean gst_asf_demux_sink_event (GstPad * pad, GstEvent * event); static GstFlowReturn gst_asf_demux_process_object (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size); GST_BOILERPLATE (GstASFDemux, gst_asf_demux, GstElement, GST_TYPE_ELEMENT); static void gst_asf_demux_base_init (gpointer g_class) { GstElementClass *element_class = GST_ELEMENT_CLASS (g_class); static GstElementDetails gst_asf_demux_details = { "ASF Demuxer", "Codec/Demuxer", "Demultiplexes ASF Streams", "Owen Fraser-Green " }; gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&audio_src_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&video_src_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_asf_demux_sink_template)); gst_element_class_set_details (element_class, &gst_asf_demux_details); GST_DEBUG_CATEGORY_INIT (asf_debug, "asfdemux", 0, "asf demuxer element"); } static void gst_asf_demux_class_init (GstASFDemuxClass * klass) { GstElementClass *gstelement_class; gstelement_class = (GstElementClass *) klass; gstelement_class->change_state = GST_DEBUG_FUNCPTR (gst_asf_demux_change_state); gstelement_class->send_event = GST_DEBUG_FUNCPTR (gst_asf_demux_element_send_event); } static void gst_asf_demux_init (GstASFDemux * demux, GstASFDemuxClass * klass) { demux->sinkpad = gst_pad_new_from_template (gst_static_pad_template_get (&gst_asf_demux_sink_template), "sink"); gst_pad_set_chain_function (demux->sinkpad, GST_DEBUG_FUNCPTR (gst_asf_demux_chain)); gst_pad_set_event_function (demux->sinkpad, GST_DEBUG_FUNCPTR (gst_asf_demux_sink_event)); gst_element_add_pad (GST_ELEMENT (demux), demux->sinkpad); /* We should zero everything to be on the safe side */ demux->num_audio_streams = 0; demux->num_video_streams = 0; demux->num_streams = 0; demux->taglist = NULL; demux->state = GST_ASF_DEMUX_STATE_HEADER; } static gboolean gst_asf_demux_sink_event (GstPad * pad, GstEvent * event) { GstASFDemux *demux; gboolean ret = TRUE; demux = GST_ASF_DEMUX (gst_pad_get_parent (pad)); GST_LOG_OBJECT (demux, "handling %s event", GST_EVENT_TYPE_NAME (event)); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_NEWSEGMENT:{ GstFormat newsegment_format; gint64 newsegment_start; guint n; gst_event_parse_new_segment (event, NULL, NULL, &newsegment_format, &newsegment_start, NULL, NULL); g_assert (newsegment_format == GST_FORMAT_BYTES); g_assert (newsegment_start >= 0); GST_OBJECT_LOCK (demux); demux->pts = 0; demux->bytes_needed = 0; demux->next_byte_offset = newsegment_start; gst_adapter_clear (demux->adapter); for (n = 0; n < demux->num_streams; n++) { if (demux->stream[n].frag_offset > 0) { gst_buffer_unref (demux->stream[n].payload); demux->stream[n].frag_offset = 0; } if (demux->stream[n].cache) { gst_buffer_unref (demux->stream[n].cache); } demux->stream[n].need_newsegment = TRUE; demux->stream[n].last_pts = GST_CLOCK_TIME_NONE; demux->stream[n].sequence = 0; } GST_OBJECT_UNLOCK (demux); break; } case GST_EVENT_FLUSH_START: case GST_EVENT_FLUSH_STOP:{ #if 0 /* just drop these events, we * send our own when seeking */ gst_event_unref (event); #endif ret = gst_pad_event_default (pad, event); break; } case GST_EVENT_EOS:{ GST_OBJECT_LOCK (demux); gst_adapter_clear (demux->adapter); demux->bytes_needed = 0; GST_OBJECT_UNLOCK (demux); gst_asf_demux_send_event_unlocked (demux, event); break; } default: ret = gst_pad_event_default (pad, event); break; } gst_object_unref (demux); return ret; } static gboolean gst_asf_demux_handle_seek_event (GstASFDemux * demux, GstEvent * event) { GstSegment segment; GstSeekFlags flags; GstSeekType cur_type, stop_type; GstFormat format; gboolean only_need_update; gboolean keyunit_sync; gboolean accurate; gboolean flush; gboolean ret = FALSE; gdouble rate; gint64 cur, stop; gint64 seek_offset; guint64 seek_packet; gst_event_parse_seek (event, &rate, &format, &flags, &cur_type, &cur, &stop_type, &stop); if (format != GST_FORMAT_TIME) { GST_LOG ("seeking is only supported in TIME format"); return FALSE; } if (rate <= 0.0) { GST_LOG ("backward playback is not supported yet"); return FALSE; } /* FIXME: this seeking code is very very broken. Do not copy * it under any circumstances, unless you want to make Wim cry */ flush = ((flags & GST_SEEK_FLAG_FLUSH) == GST_SEEK_FLAG_FLUSH); accurate = ((flags & GST_SEEK_FLAG_ACCURATE) == GST_SEEK_FLAG_ACCURATE); keyunit_sync = ((flags & GST_SEEK_FLAG_KEY_UNIT) == GST_SEEK_FLAG_KEY_UNIT); /* operating on copy of segment until we know the seek worked */ GST_OBJECT_LOCK (demux); segment = demux->segment; GST_OBJECT_UNLOCK (demux); gst_segment_set_seek (&segment, rate, format, flags, cur_type, cur, stop_type, stop, &only_need_update); GST_DEBUG ("trying to seek to time %" GST_TIME_FORMAT, GST_TIME_ARGS (segment.start)); if (demux->packet_size > 0) { gint64 seek_time = segment.start; /* Hackety hack, this sucks. We just seek to an earlier position * and let the sinks throw away the stuff before the segment start */ if (flush && (accurate || keyunit_sync)) { seek_time -= 5 * GST_SECOND; if (seek_time < 0) seek_time = 0; } seek_packet = demux->num_packets * seek_time / demux->play_time; if (seek_packet > demux->num_packets) seek_packet = demux->num_packets; seek_offset = seek_packet * demux->packet_size + demux->data_offset; /* demux->next_byte_offset will be set via newsegment event */ } else { /* FIXME */ g_message ("IMPLEMENT ME: seeking for packet_size == 0 (asfdemux)"); ret = FALSE; goto done; } GST_LOG ("seeking to byte offset %" G_GINT64_FORMAT, seek_offset); ret = gst_pad_push_event (demux->sinkpad, gst_event_new_seek (1.0, GST_FORMAT_BYTES, flags | GST_SEEK_FLAG_ACCURATE, GST_SEEK_TYPE_SET, seek_offset, GST_SEEK_TYPE_NONE, -1)); if (ret == FALSE) { GST_WARNING ("upstream element failed to seek!"); goto done; } GST_OBJECT_LOCK (demux); demux->segment = segment; demux->packet = seek_packet; GST_OBJECT_UNLOCK (demux); done: return ret; } static gboolean gst_asf_demux_handle_src_event (GstPad * pad, GstEvent * event) { GstASFDemux *demux; gboolean ret; demux = GST_ASF_DEMUX (gst_pad_get_parent (pad)); GST_DEBUG ("handling %s event on source pad %s", GST_EVENT_TYPE_NAME (event), GST_PAD_NAME (pad)); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_SEEK: ret = gst_asf_demux_handle_seek_event (demux, event); gst_event_unref (event); break; default: ret = gst_pad_event_default (pad, event); break; } gst_object_unref (demux); return ret; } static gint64 gst_asf_demux_get_current_offset (GstASFDemux * demux, guint8 * cur_data) { guint64 ret; if (demux->next_byte_offset == GST_BUFFER_OFFSET_NONE) return GST_BUFFER_OFFSET_NONE; ret = demux->next_byte_offset - gst_adapter_available (demux->adapter); if (cur_data) { guint8 *start = (guint8 *) gst_adapter_peek (demux->adapter, 1); g_assert (cur_data > start); ret += cur_data - start; } return ret; } static GstFlowReturn gst_asf_demux_chain (GstPad * pad, GstBuffer * buf) { GstFlowReturn ret = GST_FLOW_OK; GstASFDemux *demux; demux = GST_ASF_DEMUX (gst_pad_get_parent (pad)); /* GST_DEBUG ("====================== chain ================="); */ GST_DEBUG ("received buffer: size=%u, offset=%" G_GINT64_FORMAT, GST_BUFFER_SIZE (buf), GST_BUFFER_OFFSET (buf)); /* So we can always calculate the current byte offset ... */ if (GST_BUFFER_OFFSET (buf) != GST_BUFFER_OFFSET_NONE) demux->next_byte_offset = GST_BUFFER_OFFSET (buf) + GST_BUFFER_SIZE (buf); else demux->next_byte_offset = GST_BUFFER_OFFSET_NONE; gst_adapter_push (demux->adapter, buf); buf = NULL; /* adapter took ownership */ /* If we know the minimum number of bytes required * to do further processing from last time, check here * and save us some unnecessary repeated parsing */ if (demux->bytes_needed > 0) { guint avail; avail = gst_adapter_available (demux->adapter); GST_DEBUG ("bytes_needed=%u, available=%u", demux->bytes_needed, avail); if (avail < demux->bytes_needed) goto done; } demux->bytes_needed = 0; /* Parse until we need more data, get an error, or are done */ do { GST_DEBUG ("current offset = %" G_GINT64_FORMAT, gst_asf_demux_get_current_offset (demux, NULL)); ret = gst_asf_demux_parse_data (demux); } while (ret == GST_FLOW_OK); if (ret == ASF_FLOW_NEED_MORE_DATA) { GST_DEBUG ("waiting for more data, %u bytes needed and only %u available", demux->bytes_needed, gst_adapter_available (demux->adapter)); ret = GST_FLOW_OK; goto done; } GST_DEBUG ("parse_data returned %s", gst_flow_get_name (ret)); done: gst_object_unref (demux); g_assert (ret != ASF_FLOW_NEED_MORE_DATA); /* internal only */ return ret; } static inline gboolean gst_asf_demux_skip_bytes (guint num_bytes, guint8 ** p_data, guint64 * p_size) { if (*p_size < num_bytes) return FALSE; *p_data += num_bytes; *p_size -= num_bytes; return TRUE; } static inline guint32 gst_asf_demux_identify_guid (GstASFDemux * demux, const ASFGuidHash * guids, ASFGuid * guid) { guint32 ret; ret = gst_asf_identify_guid (guids, guid); GST_LOG ("%s 0x%08x-0x%08x-0x%08x-0x%08x", gst_asf_get_guid_nick (guids, ret), guid->v1, guid->v2, guid->v3, guid->v4); return ret; } static inline guint8 gst_asf_demux_get_uint8 (guint8 ** p_data, guint64 * p_size) { guint8 ret; g_assert (*p_size >= 1); ret = GST_READ_UINT8 (*p_data); *p_data += sizeof (guint8); *p_size -= sizeof (guint8); return ret; } static inline guint16 gst_asf_demux_get_uint16 (guint8 ** p_data, guint64 * p_size) { guint16 ret; g_assert (*p_size >= 2); ret = GST_READ_UINT16_LE (*p_data); *p_data += sizeof (guint16); *p_size -= sizeof (guint16); return ret; } static inline guint32 gst_asf_demux_get_uint32 (guint8 ** p_data, guint64 * p_size) { guint32 ret; g_assert (*p_size >= 4); ret = GST_READ_UINT32_LE (*p_data); *p_data += sizeof (guint32); *p_size -= sizeof (guint32); return ret; } static inline guint64 gst_asf_demux_get_uint64 (guint8 ** p_data, guint64 * p_size) { guint64 ret; g_assert (*p_size >= 8); ret = GST_READ_UINT64_LE (*p_data); *p_data += sizeof (guint64); *p_size -= sizeof (guint64); return ret; } static inline guint32 gst_asf_demux_get_var_length (guint8 type, guint8 ** p_data, guint64 * p_size) { switch (type) { case 0: return 0; case 1: g_assert (*p_size >= 1); return gst_asf_demux_get_uint8 (p_data, p_size); case 2: g_assert (*p_size >= 2); return gst_asf_demux_get_uint16 (p_data, p_size); case 3: g_assert (*p_size >= 4); return gst_asf_demux_get_uint32 (p_data, p_size); default: break; } g_assert_not_reached (); } static gboolean gst_asf_demux_get_buffer (GstBuffer ** p_buf, guint num_bytes_to_read, guint8 ** p_data, guint64 * p_size) { *p_buf = NULL; if (*p_size < num_bytes_to_read) return FALSE; *p_buf = gst_buffer_new_and_alloc (num_bytes_to_read); memcpy (GST_BUFFER_DATA (*p_buf), *p_data, num_bytes_to_read); *p_data += num_bytes_to_read; *p_size -= num_bytes_to_read; return TRUE; } static gboolean gst_asf_demux_get_bytes (guint8 ** p_buf, guint num_bytes_to_read, guint8 ** p_data, guint64 * p_size) { *p_buf = NULL; if (*p_size < num_bytes_to_read) return FALSE; *p_buf = g_memdup (*p_data, num_bytes_to_read); *p_data += num_bytes_to_read; *p_size -= num_bytes_to_read; return TRUE; } static gboolean gst_asf_demux_get_string (gchar ** p_str, guint16 * p_strlen, guint8 ** p_data, guint64 * p_size) { guint16 s_length; guint8 *s; *p_str = NULL; if (*p_size < 2) return FALSE; s_length = gst_asf_demux_get_uint16 (p_data, p_size); if (p_strlen) *p_strlen = s_length; if (s_length == 0) { GST_WARNING ("zero-length string"); *p_str = g_strdup (""); return TRUE; } if (!gst_asf_demux_get_bytes (&s, s_length, p_data, p_size)) return FALSE; g_assert (s != NULL); /* just because They don't exist doesn't * mean They are not out to get you ... */ if (s[s_length - 1] != '\0') { s = g_realloc (s, s_length + 1); s[s_length] = '\0'; } *p_str = (gchar *) s; return TRUE; } static gboolean gst_asf_demux_get_guid (ASFGuid * guid, guint8 ** p_data, guint64 * p_size) { if (*p_size < 4 * sizeof (guint32)) return FALSE; guid->v1 = gst_asf_demux_get_uint32 (p_data, p_size); guid->v2 = gst_asf_demux_get_uint32 (p_data, p_size); guid->v3 = gst_asf_demux_get_uint32 (p_data, p_size); guid->v4 = gst_asf_demux_get_uint32 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_obj_file (asf_obj_file * object, guint8 ** p_data, guint64 * p_size) { if (*p_size < (16 + 8 + 8 + 8 + 8 + 8 + 8 + 4 + 4 + 4 + 4)) return FALSE; gst_asf_demux_get_guid (&object->file_id, p_data, p_size); object->file_size = gst_asf_demux_get_uint64 (p_data, p_size); object->creation_time = gst_asf_demux_get_uint64 (p_data, p_size); object->packets_count = gst_asf_demux_get_uint64 (p_data, p_size); object->play_time = gst_asf_demux_get_uint64 (p_data, p_size); object->send_time = gst_asf_demux_get_uint64 (p_data, p_size); object->preroll = gst_asf_demux_get_uint64 (p_data, p_size); object->flags = gst_asf_demux_get_uint32 (p_data, p_size); object->min_pktsize = gst_asf_demux_get_uint32 (p_data, p_size); object->max_pktsize = gst_asf_demux_get_uint32 (p_data, p_size); object->min_bitrate = gst_asf_demux_get_uint32 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_bitrate_record (asf_bitrate_record * record, guint8 ** p_data, guint64 * p_size) { if (*p_size < (2 + 4)) return FALSE; record->stream_id = gst_asf_demux_get_uint16 (p_data, p_size); record->bitrate = gst_asf_demux_get_uint32 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_obj_comment (asf_obj_comment * comment, guint8 ** p_data, guint64 * p_size) { if (*p_size < (2 + 2 + 2 + 2 + 2)) return FALSE; comment->title_length = gst_asf_demux_get_uint16 (p_data, p_size); comment->author_length = gst_asf_demux_get_uint16 (p_data, p_size); comment->copyright_length = gst_asf_demux_get_uint16 (p_data, p_size); comment->description_length = gst_asf_demux_get_uint16 (p_data, p_size); comment->rating_length = gst_asf_demux_get_uint16 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_obj_header (asf_obj_header * header, guint8 ** p_data, guint64 * p_size) { if (*p_size < (4 + 1 + 1)) return FALSE; header->num_objects = gst_asf_demux_get_uint32 (p_data, p_size); header->unknown1 = gst_asf_demux_get_uint8 (p_data, p_size); header->unknown2 = gst_asf_demux_get_uint8 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_obj_header_ext (asf_obj_header_ext * hdr_ext, guint8 ** p_data, guint64 * p_size) { if (*p_size < (16 + 2 + 4)) return FALSE; gst_asf_demux_get_guid (&hdr_ext->reserved1, p_data, p_size); hdr_ext->reserved2 = gst_asf_demux_get_uint16 (p_data, p_size); hdr_ext->data_size = gst_asf_demux_get_uint32 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_obj_stream (asf_obj_stream * stream, guint8 ** p_data, guint64 * p_size) { guint16 flags; if (*p_size < (16 + 16 + 8 + 4 + 4 + 2 + 4)) return FALSE; gst_asf_demux_get_guid (&stream->type, p_data, p_size); gst_asf_demux_get_guid (&stream->correction, p_data, p_size); stream->time_offset = gst_asf_demux_get_uint64 (p_data, p_size) * 100; stream->type_specific_size = gst_asf_demux_get_uint32 (p_data, p_size); stream->stream_specific_size = gst_asf_demux_get_uint32 (p_data, p_size); flags = gst_asf_demux_get_uint16 (p_data, p_size); stream->id = flags & 0x7f; stream->encrypted = (flags & 0x8000) << 15; stream->unknown2 = gst_asf_demux_get_uint32 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_replicated_data (asf_replicated_data * rep, guint8 ** p_data, guint64 * p_size) { if (*p_size < (4 + 4)) return FALSE; rep->object_size = gst_asf_demux_get_uint32 (p_data, p_size); rep->frag_timestamp = gst_asf_demux_get_uint32 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_obj_data (asf_obj_data * object, guint8 ** p_data, guint64 * p_size) { if (*p_size < (16 + 8 + 1 + 1)) return FALSE; gst_asf_demux_get_guid (&object->file_id, p_data, p_size); object->packets = gst_asf_demux_get_uint64 (p_data, p_size); object->unknown1 = gst_asf_demux_get_uint8 (p_data, p_size); /* object->unknown2 = gst_asf_demux_get_uint8 (p_data, p_size); */ object->correction = gst_asf_demux_get_uint8 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_obj_data_correction (asf_obj_data_correction * object, guint8 ** p_data, guint64 * p_size) { if (*p_size < (1 + 1)) return FALSE; object->type = gst_asf_demux_get_uint8 (p_data, p_size); object->cycle = gst_asf_demux_get_uint8 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_stream_audio (asf_stream_audio * audio, guint8 ** p_data, guint64 * p_size) { if (*p_size < (2 + 2 + 4 + 4 + 2 + 2 + 2)) return FALSE; /* WAVEFORMATEX Structure */ audio->codec_tag = gst_asf_demux_get_uint16 (p_data, p_size); audio->channels = gst_asf_demux_get_uint16 (p_data, p_size); audio->sample_rate = gst_asf_demux_get_uint32 (p_data, p_size); audio->byte_rate = gst_asf_demux_get_uint32 (p_data, p_size); audio->block_align = gst_asf_demux_get_uint16 (p_data, p_size); audio->word_size = gst_asf_demux_get_uint16 (p_data, p_size); /* Codec specific data size */ audio->size = gst_asf_demux_get_uint16 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_stream_correction (asf_stream_correction * object, guint8 ** p_data, guint64 * p_size) { if (*p_size < (1 + 2 + 2 + 2 + 1)) return FALSE; object->span = gst_asf_demux_get_uint8 (p_data, p_size); object->packet_size = gst_asf_demux_get_uint16 (p_data, p_size); object->chunk_size = gst_asf_demux_get_uint16 (p_data, p_size); object->data_size = gst_asf_demux_get_uint16 (p_data, p_size); object->silence_data = gst_asf_demux_get_uint8 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_stream_video (asf_stream_video * video, guint8 ** p_data, guint64 * p_size) { if (*p_size < (4 + 4 + 1 + 2)) return FALSE; video->width = gst_asf_demux_get_uint32 (p_data, p_size); video->height = gst_asf_demux_get_uint32 (p_data, p_size); video->unknown = gst_asf_demux_get_uint8 (p_data, p_size); video->size = gst_asf_demux_get_uint16 (p_data, p_size); return TRUE; } static gboolean gst_asf_demux_get_stream_video_format (asf_stream_video_format * fmt, guint8 ** p_data, guint64 * p_size) { if (*p_size < (4 + 4 + 4 + 2 + 2 + 4 + 4 + 4 + 4 + 4 + 4)) return FALSE; fmt->size = gst_asf_demux_get_uint32 (p_data, p_size); fmt->width = gst_asf_demux_get_uint32 (p_data, p_size); fmt->height = gst_asf_demux_get_uint32 (p_data, p_size); fmt->planes = gst_asf_demux_get_uint16 (p_data, p_size); fmt->depth = gst_asf_demux_get_uint16 (p_data, p_size); fmt->tag = gst_asf_demux_get_uint32 (p_data, p_size); fmt->image_size = gst_asf_demux_get_uint32 (p_data, p_size); fmt->xpels_meter = gst_asf_demux_get_uint32 (p_data, p_size); fmt->ypels_meter = gst_asf_demux_get_uint32 (p_data, p_size); fmt->num_colors = gst_asf_demux_get_uint32 (p_data, p_size); fmt->imp_colors = gst_asf_demux_get_uint32 (p_data, p_size); return TRUE; } static asf_stream_context * gst_asf_demux_get_stream (GstASFDemux * demux, guint16 id) { guint8 i; asf_stream_context *stream; for (i = 0; i < demux->num_streams; i++) { stream = &demux->stream[i]; if (stream->id == id) { /* We've found the one with the matching id */ return &demux->stream[i]; } } /* Base case if we haven't found one at all */ GST_WARNING ("Segment found for undefined stream: (%d)", id); return NULL; } static asf_obj_ext_stream_properties * gst_asf_demux_get_ext_stream_props_for_stream (GstASFDemux * demux, gint id) { GSList *l; for (l = demux->ext_stream_props; l != NULL; l = l->next) { asf_obj_ext_stream_properties *esp; esp = (asf_obj_ext_stream_properties *) l->data; if (esp->stream_num == id) return esp; } return NULL; } static void gst_asf_demux_setup_pad (GstASFDemux * demux, GstPad * src_pad, GstCaps * caps, guint16 id, gboolean is_video, GstTagList * tags) { asf_stream_context *stream; gst_pad_use_fixed_caps (src_pad); gst_pad_set_caps (src_pad, caps); gst_pad_set_event_function (src_pad, GST_DEBUG_FUNCPTR (gst_asf_demux_handle_src_event)); gst_pad_set_query_type_function (src_pad, GST_DEBUG_FUNCPTR (gst_asf_demux_get_src_query_types)); gst_pad_set_query_function (src_pad, GST_DEBUG_FUNCPTR (gst_asf_demux_handle_src_query)); stream = &demux->stream[demux->num_streams]; stream->caps = caps; stream->pad = src_pad; stream->id = id; stream->frag_offset = 0; stream->sequence = 0; stream->delay = 0; stream->first_pts = GST_CLOCK_TIME_NONE; stream->last_pts = GST_CLOCK_TIME_NONE; stream->fps_known = !is_video; /* bit hacky for audio */ stream->is_video = is_video; stream->need_newsegment = TRUE; stream->pending_tags = tags; gst_pad_set_element_private (src_pad, stream); GST_INFO ("Adding pad %s for stream %u with caps %" GST_PTR_FORMAT, GST_PAD_NAME (src_pad), demux->num_streams, caps); ++demux->num_streams; gst_element_add_pad (GST_ELEMENT (demux), src_pad); } static void gst_asf_demux_add_audio_stream (GstASFDemux * demux, asf_stream_audio * audio, guint16 id, guint8 ** p_data, guint64 * p_size) { asf_obj_ext_stream_properties *ext_props; GstTagList *tags = NULL; GstBuffer *extradata = NULL; GstPad *src_pad; GstCaps *caps; guint16 size_left = 0; gchar *codec_name = NULL; gchar *name = NULL; size_left = audio->size; /* Create the audio pad */ name = g_strdup_printf ("audio_%02d", demux->num_audio_streams); src_pad = gst_pad_new_from_static_template (&audio_src_template, name); g_free (name); /* Swallow up any left over data and set up the * standard properties from the header info */ if (size_left) { GST_WARNING ("Audio header contains %d bytes of " "codec specific data", size_left); gst_asf_demux_get_buffer (&extradata, size_left, p_data, p_size); } /* asf_stream_audio is the same as gst_riff_strf_auds, but with an * additional two bytes indicating extradata. */ caps = gst_riff_create_audio_caps (audio->codec_tag, NULL, (gst_riff_strf_auds *) audio, extradata, NULL, &codec_name); if (caps == NULL) { caps = gst_caps_new_simple ("audio/x-asf-unknown", "codec_id", G_TYPE_INT, (gint) audio->codec_tag, NULL); } /* Informing about that audio format we just added */ if (codec_name) { tags = gst_tag_list_new (); gst_tag_list_add (tags, GST_TAG_MERGE_APPEND, GST_TAG_AUDIO_CODEC, codec_name, NULL); g_free (codec_name); } /* add language info if we have it */ ext_props = gst_asf_demux_get_ext_stream_props_for_stream (demux, id); if (ext_props && ext_props->lang_idx < demux->num_languages) { if (tags == NULL) tags = gst_tag_list_new (); gst_tag_list_add (tags, GST_TAG_MERGE_APPEND, GST_TAG_LANGUAGE_CODE, demux->languages[ext_props->lang_idx], NULL); } if (extradata) gst_buffer_unref (extradata); GST_INFO ("Adding audio stream %u codec %u (0x%04x), tags=%" GST_PTR_FORMAT, demux->num_video_streams, audio->codec_tag, audio->codec_tag, tags); ++demux->num_audio_streams; gst_asf_demux_setup_pad (demux, src_pad, caps, id, FALSE, tags); } static void gst_asf_demux_add_video_stream (GstASFDemux * demux, asf_stream_video_format * video, guint16 id, guint8 ** p_data, guint64 * p_size) { GstTagList *tags = NULL; GstBuffer *extradata = NULL; GstPad *src_pad; GstCaps *caps; gchar *name = NULL; gchar *codec_name = NULL; gint size_left = video->size - 40; /* Create the video pad */ name = g_strdup_printf ("video_%02d", demux->num_video_streams); src_pad = gst_pad_new_from_static_template (&video_src_template, name); g_free (name); /* Now try some gstreamer formatted MIME types (from gst_avi_demux_strf_vids) */ if (size_left) { GST_LOG ("Video header has %d bytes of codec specific data", size_left); gst_asf_demux_get_buffer (&extradata, size_left, p_data, p_size); } GST_DEBUG ("video codec %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (video->tag)); /* yes, asf_stream_video_format and gst_riff_strf_vids are the same */ caps = gst_riff_create_video_caps (video->tag, NULL, (gst_riff_strf_vids *) video, extradata, NULL, &codec_name); if (caps == NULL) { caps = gst_caps_new_simple ("video/x-asf-unknown", "fourcc", GST_TYPE_FOURCC, video->tag, NULL); } if (codec_name) { tags = gst_tag_list_new (); gst_tag_list_add (tags, GST_TAG_MERGE_APPEND, GST_TAG_VIDEO_CODEC, codec_name, NULL); g_free (codec_name); } if (extradata) gst_buffer_unref (extradata); GST_INFO ("Adding video stream %u codec %" GST_FOURCC_FORMAT " (0x%08x)", demux->num_video_streams, GST_FOURCC_ARGS (video->tag), video->tag); gst_caps_set_simple (caps, "framerate", GST_TYPE_FRACTION, 25, 1, NULL); ++demux->num_video_streams; gst_asf_demux_setup_pad (demux, src_pad, caps, id, TRUE, tags); } static GstFlowReturn gst_asf_demux_process_stream (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { asf_obj_stream object; guint32 stream_id; guint32 correction; guint8 *obj_data_start = *p_data; /* Get the rest of the header's header */ if (!gst_asf_demux_get_obj_stream (&object, p_data, p_size)) goto not_enough_data; GST_DEBUG ("Found stream #%u", object.id); /* Identify the stream type */ stream_id = gst_asf_demux_identify_guid (demux, asf_stream_guids, &object.type); correction = gst_asf_demux_identify_guid (demux, asf_correction_guids, &object.correction); switch (stream_id) { case ASF_STREAM_AUDIO:{ asf_stream_correction correction_object; asf_stream_audio audio_object; if (!gst_asf_demux_get_stream_audio (&audio_object, p_data, p_size)) goto not_enough_data; GST_INFO ("Object is an audio stream with %u bytes of additional data", audio_object.size); gst_asf_demux_add_audio_stream (demux, &audio_object, object.id, p_data, p_size); switch (correction) { case ASF_CORRECTION_ON: GST_INFO ("Using error correction"); if (!gst_asf_demux_get_stream_correction (&correction_object, p_data, p_size)) { goto not_enough_data; } demux->span = correction_object.span; GST_DEBUG ("Descrambling: ps:%d cs:%d ds:%d s:%d sd:%d", correction_object.packet_size, correction_object.chunk_size, correction_object.data_size, (guint) correction_object.span, (guint) correction_object.silence_data); if (demux->span > 1) { if (!correction_object.chunk_size || ((correction_object.packet_size / correction_object.chunk_size) <= 1)) { /* Disable descrambling */ demux->span = 0; } else { /* FIXME: this else branch was added for * weird_al_yankovic - the saga begins.asf */ demux->ds_packet_size = correction_object.packet_size; demux->ds_chunk_size = correction_object.chunk_size; } } else { /* Descambling is enabled */ demux->ds_packet_size = correction_object.packet_size; demux->ds_chunk_size = correction_object.chunk_size; } #if 0 /* Now skip the rest of the silence data */ if (correction_object.data_size > 1) gst_bytestream_flush (demux->bs, correction_object.data_size - 1); #else /* FIXME: CHECKME. And why -1? */ if (correction_object.data_size > 1) { if (!gst_asf_demux_skip_bytes (correction_object.data_size - 1, p_data, p_size)) { goto not_enough_data; } } #endif break; case ASF_CORRECTION_OFF: GST_INFO ("Error correction off"); #if 0 /* gst_bytestream_flush (demux->bs, object.stream_specific_size); */ #else /* FIXME: CHECKME */ if (!gst_asf_demux_skip_bytes (object.stream_specific_size, p_data, p_size)) { goto not_enough_data; } #endif break; default: GST_ELEMENT_ERROR (demux, STREAM, DEMUX, (NULL), ("Audio stream using unknown error correction")); return GST_FLOW_ERROR; } break; } case ASF_STREAM_VIDEO:{ asf_stream_video_format video_format_object; asf_stream_video video_object; guint16 size; if (!gst_asf_demux_get_stream_video (&video_object, p_data, p_size)) goto not_enough_data; size = video_object.size - 40; /* Byte order gets offset by single byte */ GST_INFO ("object is a video stream with %u bytes of " "additional data", size); if (!gst_asf_demux_get_stream_video_format (&video_format_object, p_data, p_size)) { goto not_enough_data; } gst_asf_demux_add_video_stream (demux, &video_format_object, object.id, p_data, p_size); break; } default: GST_WARNING_OBJECT (demux, "Unknown asf stream (id %08x)", (guint) stream_id); break; } return GST_FLOW_OK; not_enough_data: { /* avoid compiler warning when disabling logging at compile time */ obj_data_start = NULL; GST_WARNING ("Unexpected end of data parsing stream object"); GST_DEBUG ("object data offset: %u, bytes left to parse: %u", (guint) (*p_data - obj_data_start), (guint) * p_size); return ASF_FLOW_NEED_MORE_DATA; } } static const gchar * gst_asf_demux_get_gst_tag_from_tag_name (const gchar * name_utf16le, gsize name_len) { const struct { const gchar *asf_name; const gchar *gst_name; } tags[] = { { "WM/Genre", GST_TAG_GENRE}, { "WM/AlbumTitle", GST_TAG_ALBUM}, { "WM/AlbumArtist", GST_TAG_ARTIST}, { "WM/Track", GST_TAG_TRACK_NUMBER}, { "WM/Year", GST_TAG_DATE} /* { "WM/Composer", GST_TAG_COMPOSER } */ }; gchar *name_utf8; gsize in, out; guint i; /* convert name to UTF-8 */ name_utf8 = g_convert (name_utf16le, name_len, "UTF-8", "UTF-16LE", &in, &out, NULL); if (name_utf8 == NULL) { GST_WARNING ("Failed to convert name to UTF8, skipping"); return NULL; } for (i = 0; i < G_N_ELEMENTS (tags); ++i) { if (strncmp (tags[i].asf_name, name_utf8, out) == 0) { GST_LOG ("map tagname '%s' -> '%s'", name_utf8, tags[i].gst_name); g_free (name_utf8); return tags[i].gst_name; } } GST_LOG ("unhandled tagname '%s'", name_utf8); g_free (name_utf8); return NULL; } /* gst_asf_demux_commit_taglist() takes ownership of taglist! */ static void gst_asf_demux_commit_taglist (GstASFDemux * demux, GstTagList * taglist) { GST_DEBUG ("Committing tags: %" GST_PTR_FORMAT, taglist); gst_element_found_tags (GST_ELEMENT (demux), gst_tag_list_copy (taglist)); /* save internally */ if (!demux->taglist) demux->taglist = taglist; else { GstTagList *t; t = gst_tag_list_merge (demux->taglist, taglist, GST_TAG_MERGE_APPEND); gst_tag_list_free (demux->taglist); gst_tag_list_free (taglist); demux->taglist = t; } } #define ASF_DEMUX_DATA_TYPE_UTF16LE_STRING 0 #define ASF_DEMUX_DATA_TYPE_DWORD 3 /* Extended Content Description Object */ static GstFlowReturn gst_asf_demux_process_ext_content_desc (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { /* Other known (and unused) 'text/unicode' metadata available : * * WM/Lyrics = * WM/MediaPrimaryClassID = {D1607DBC-E323-4BE2-86A1-48A42A28441E} * WMFSDKVersion = 9.00.00.2980 * WMFSDKNeeded = 0.0.0.0000 * WM/UniqueFileIdentifier = AMGa_id=R 15334;AMGp_id=P 5149;AMGt_id=T 2324984 * WM/Publisher = 4AD * WM/Provider = AMG * WM/ProviderRating = 8 * WM/ProviderStyle = Rock (similar to WM/Genre) * WM/GenreID (similar to WM/Genre) * WM/TrackNumber (same as WM/Track but as a string) * * Other known (and unused) 'non-text' metadata available : * * WM/EncodingTime * WM/MCDI * IsVBR */ GstTagList *taglist; guint16 blockcount, i; guint8 *obj_data_start = *p_data; GST_INFO ("object is an extended content description"); taglist = gst_tag_list_new (); /* Content Descriptor Count */ if (*p_size < 2) goto not_enough_data; blockcount = gst_asf_demux_get_uint16 (p_data, p_size); for (i = 1; i <= blockcount; ++i) { const gchar *gst_tag_name; guint16 datatype; guint16 value_len; guint16 name_len; GValue tag_value = { 0, }; gsize in, out; gchar *name; gchar *value; /* Descriptor */ if (!gst_asf_demux_get_string (&name, &name_len, p_data, p_size)) goto not_enough_data; if (*p_size < 2) goto not_enough_data; /* Descriptor Value Data Type */ datatype = gst_asf_demux_get_uint16 (p_data, p_size); /* Descriptor Value (not really a string, but same thing reading-wise) */ if (!gst_asf_demux_get_string (&value, &value_len, p_data, p_size)) goto not_enough_data; gst_tag_name = gst_asf_demux_get_gst_tag_from_tag_name (name, name_len); if (gst_tag_name != NULL) { switch (datatype) { case ASF_DEMUX_DATA_TYPE_UTF16LE_STRING:{ gchar *value_utf8; value_utf8 = g_convert (value, value_len, "UTF-8", "UTF-16LE", &in, &out, NULL); /* get rid of tags with empty value */ if (value_utf8 != NULL && *value_utf8 != '\0') { value_utf8[out] = '\0'; if (strcmp (gst_tag_name, GST_TAG_DATE) == 0) { guint year = atoi (value_utf8); if (year > 0) { GDate *date = g_date_new_dmy (1, 1, year); g_value_init (&tag_value, GST_TYPE_DATE); gst_value_set_date (&tag_value, date); g_date_free (date); } } else { g_value_init (&tag_value, G_TYPE_STRING); g_value_set_string (&tag_value, value_utf8); } } else if (value_utf8 == NULL) { GST_WARNING ("Failed to convert string value to UTF8, skipping"); } else { GST_DEBUG ("Skipping empty string value for %s", gst_tag_name); } g_free (value_utf8); break; } case ASF_DEMUX_DATA_TYPE_DWORD:{ /* this is the track number */ g_value_init (&tag_value, G_TYPE_UINT); g_value_set_uint (&tag_value, (guint) GST_READ_UINT32_LE (value)); break; } default:{ GST_DEBUG ("Skipping tag %s of type %d", gst_tag_name, datatype); break; } } if (G_IS_VALUE (&tag_value)) { gst_tag_list_add_values (taglist, GST_TAG_MERGE_APPEND, gst_tag_name, &tag_value, NULL); g_value_unset (&tag_value); } } g_free (name); g_free (value); } if (gst_structure_n_fields (GST_STRUCTURE (taglist)) > 0) { gst_asf_demux_commit_taglist (demux, taglist); } else { gst_tag_list_free (taglist); } return GST_FLOW_OK; /* Errors */ not_enough_data: { /* avoid compiler warning when disabling logging at compile time */ obj_data_start = NULL; GST_WARNING ("Unexpected end of data parsing stream object"); GST_DEBUG ("object data offset: %u, bytes left to parse: %u", (guint) (*p_data - obj_data_start), (guint) * p_size); gst_tag_list_free (taglist); return ASF_FLOW_NEED_MORE_DATA; } } #define ASF_DEMUX_OBJECT_HEADER_SIZE (16+8) static gboolean gst_asf_demux_get_object_header (GstASFDemux * demux, guint32 * obj_id, guint64 * obj_size, guint8 ** p_data, guint64 * p_size) { ASFGuid guid; if (*p_size < ASF_DEMUX_OBJECT_HEADER_SIZE) return FALSE; gst_asf_demux_get_guid (&guid, p_data, p_size); *obj_id = gst_asf_demux_identify_guid (demux, asf_object_guids, &guid); *obj_size = gst_asf_demux_get_uint64 (p_data, p_size); if (*obj_id == ASF_OBJ_UNDEFINED) { GST_WARNING ("Unknown object %08x-%08x-%08x-%08x", guid.v1, guid.v2, guid.v3, guid.v4); } return TRUE; } static GstFlowReturn gst_asf_demux_process_data (GstASFDemux * demux, guint64 object_size, guint8 ** p_data, guint64 * p_size) { asf_obj_data data_object; /* Get the rest of the header */ if (!gst_asf_demux_get_obj_data (&data_object, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; GST_INFO ("object is data with %" G_GUINT64_FORMAT " packets", data_object.packets); gst_element_no_more_pads (GST_ELEMENT (demux)); demux->state = GST_ASF_DEMUX_STATE_DATA; demux->packet = 0; demux->num_packets = data_object.packets; /* minus object header and data object header */ demux->data_size = object_size - ASF_DEMUX_OBJECT_HEADER_SIZE - (16 + 8 + 1 + 1); demux->data_offset = gst_asf_demux_get_current_offset (demux, *p_data); GST_LOG ("data_offset=%" G_GINT64_FORMAT ", data_size=%" G_GINT64_FORMAT, demux->data_offset, demux->data_size); return GST_FLOW_OK; } static GstFlowReturn gst_asf_demux_process_header (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { GstFlowReturn ret = GST_FLOW_OK; asf_obj_header object; guint32 i; /* Get the rest of the header's header */ if (!gst_asf_demux_get_obj_header (&object, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; GST_INFO ("object is a header with %u parts", object.num_objects); /* Loop through the header's objects, processing those */ for (i = 0; i < object.num_objects; ++i) { GST_DEBUG ("reading header part %u: offset=0x%" G_GINT64_MODIFIER "x", i, gst_asf_demux_get_current_offset (demux, *p_data)); ret = gst_asf_demux_process_object (demux, p_data, p_size); if (ret != GST_FLOW_OK) { GST_WARNING ("process_object returned %s", gst_asf_get_flow_name (ret)); break; } } return ret; } static GstFlowReturn gst_asf_demux_process_file (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { asf_obj_file object; /* Get the rest of the header's header */ if (!gst_asf_demux_get_obj_file (&object, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; if (object.min_pktsize == object.max_pktsize) { demux->packet_size = object.max_pktsize; } else { demux->packet_size = (guint32) - 1; GST_WARNING ("Non-const packet size, seeking disabled"); } /* FIXME: do we need object.send_time as well? what is it? */ demux->play_time = (guint64) object.play_time * (GST_SECOND / 10000000); demux->preroll = object.preroll; GST_DEBUG_OBJECT (demux, "play_time %" GST_TIME_FORMAT " preroll %" GST_TIME_FORMAT, GST_TIME_ARGS (demux->play_time), GST_TIME_ARGS (demux->preroll)); gst_segment_set_duration (&demux->segment, GST_FORMAT_TIME, demux->play_time); GST_INFO ("object is a file with %" G_GUINT64_FORMAT " data packets", object.packets_count); GST_INFO ("preroll = %" G_GUINT64_FORMAT, demux->preroll); return GST_FLOW_OK; } /* Content Description Object */ static GstFlowReturn gst_asf_demux_process_comment (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { struct { const gchar *gst_tag; guint16 val_length; gchar *val_utf8; } tags[5] = { { GST_TAG_TITLE, 0, NULL}, { GST_TAG_ARTIST, 0, NULL}, { GST_TAG_COPYRIGHT, 0, NULL}, { GST_TAG_COMMENT, 0, NULL}, { NULL, 0, NULL} /* what GST_TAG to use here? */ }; asf_obj_comment object; GstTagList *taglist; GValue value = { 0 }; gsize in, out; gint i; GST_INFO ("object is a comment"); /* Get the rest of the comment's header */ if (!gst_asf_demux_get_obj_comment (&object, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; GST_DEBUG ("Comment lengths: title=%d author=%d copyright=%d " "description=%d rating=%d", object.title_length, object.author_length, object.copyright_length, object.description_length, object.rating_length); tags[0].val_length = object.title_length; tags[1].val_length = object.author_length, tags[2].val_length = object.copyright_length; tags[3].val_length = object.description_length; tags[4].val_length = object.rating_length; for (i = 0; i < G_N_ELEMENTS (tags); ++i) { if (*p_size < tags[i].val_length) goto not_enough_data; /* might be just '/0', '/0'... */ if (tags[i].val_length > 2 && tags[i].val_length % 2 == 0) { /* convert to UTF-8 */ tags[i].val_utf8 = g_convert ((gchar *) * p_data, tags[i].val_length, "UTF-8", "UTF-16LE", &in, &out, NULL); } *p_data += tags[i].val_length; *p_size -= tags[i].val_length; } /* parse metadata into taglist */ taglist = gst_tag_list_new (); g_value_init (&value, G_TYPE_STRING); for (i = 0; i < G_N_ELEMENTS (tags); ++i) { if (tags[i].val_utf8 && strlen (tags[i].val_utf8) > 0 && tags[i].gst_tag) { g_value_set_string (&value, tags[i].val_utf8); gst_tag_list_add_values (taglist, GST_TAG_MERGE_APPEND, tags[i].gst_tag, &value, NULL); } } g_value_unset (&value); if (gst_structure_n_fields (GST_STRUCTURE (taglist)) > 0) { gst_asf_demux_commit_taglist (demux, taglist); } else { gst_tag_list_free (taglist); } for (i = 0; i < G_N_ELEMENTS (tags); ++i) g_free (tags[i].val_utf8); return GST_FLOW_OK; not_enough_data: { GST_WARNING ("unexpectedly short of data while processing " "comment tag section %s, skipping comment tag", (i < G_N_ELEMENTS (tags)) ? tags[i].gst_tag : "NONE"); for (i = 0; i < G_N_ELEMENTS (tags); i++) g_free (tags[i].val_utf8); return GST_FLOW_OK; } } static GstFlowReturn gst_asf_demux_process_bitrate_props_object (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { guint16 num_streams, i; guint8 stream_id; if (*p_size < 2) return ASF_FLOW_NEED_MORE_DATA; num_streams = gst_asf_demux_get_uint16 (p_data, p_size); GST_INFO ("object is a bitrate properties object with %u streams", num_streams); for (i = 0; i < num_streams; ++i) { asf_bitrate_record bitrate_record; if (!gst_asf_demux_get_bitrate_record (&bitrate_record, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; stream_id = bitrate_record.stream_id; if (bitrate_record.stream_id < GST_ASF_DEMUX_NUM_STREAM_IDS) { demux->bitrate[stream_id] = bitrate_record.bitrate; GST_DEBUG ("bitrate[%u] = %u", stream_id, bitrate_record.bitrate); } else { GST_WARNING ("stream id %u is too large", stream_id); } } return GST_FLOW_OK; } static GstFlowReturn gst_asf_demux_process_header_ext (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { GstFlowReturn ret = GST_FLOW_OK; asf_obj_header_ext object; guint64 target_size; /* Get the rest of the header's header */ if (!gst_asf_demux_get_obj_header_ext (&object, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; GST_INFO ("object is an extended header with a size of %u bytes", object.data_size); /* FIXME: does data_size include the rest of the header that we have read? */ if (*p_size < object.data_size) return ASF_FLOW_NEED_MORE_DATA; target_size = *p_size - object.data_size; while (*p_size > target_size && ret == GST_FLOW_OK) { ret = gst_asf_demux_process_object (demux, p_data, p_size); if (ret != GST_FLOW_OK) break; } return ret; } static GstFlowReturn gst_asf_demux_process_language_list (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { guint i; if (*p_size < 2) return ASF_FLOW_NEED_MORE_DATA; if (demux->languages) { GST_WARNING ("More than one LANGUAGE_LIST object in stream"); g_strfreev (demux->languages); demux->languages = NULL; demux->num_languages = 0; } demux->num_languages = gst_asf_demux_get_uint16 (p_data, p_size); GST_LOG ("%u languages:", demux->num_languages); demux->languages = g_new0 (gchar *, demux->num_languages + 1); for (i = 0; i < demux->num_languages; ++i) { guint8 len, *data = NULL; len = gst_asf_demux_get_uint8 (p_data, p_size); if (gst_asf_demux_get_bytes (&data, len, p_data, p_size)) { gchar *utf8; utf8 = g_convert ((gchar *) data, len, "UTF-8", "UTF-16LE", NULL, NULL, NULL); /* truncate "en-us" etc. to just "en" */ if (utf8 && strlen (utf8) >= 5 && (utf8[2] == '-' || utf8[2] == '_')) { utf8[2] = '\0'; } GST_DEBUG ("[%u] %s", i, GST_STR_NULL (utf8)); demux->languages[i] = utf8; g_free (data); } } return GST_FLOW_OK; } static GstFlowReturn gst_asf_demux_process_ext_stream_props (GstASFDemux * demux, guint obj_size, guint8 ** p_data, guint64 * p_size) { asf_obj_ext_stream_properties esp = { 0, }; guint16 stream_name_count; guint16 payload_ext_sys_count; guint64 len; guint32 obj_id; guint8 *data; guint8 *data_start = *p_data; guint i; if (*p_size < 88) return ASF_FLOW_NEED_MORE_DATA; esp.start_time = gst_asf_demux_get_uint64 (p_data, p_size); esp.end_time = gst_asf_demux_get_uint64 (p_data, p_size); esp.data_bitrate = gst_asf_demux_get_uint32 (p_data, p_size); esp.buffer_size = gst_asf_demux_get_uint32 (p_data, p_size); esp.intial_buf_fullness = gst_asf_demux_get_uint32 (p_data, p_size); esp.data_bitrate2 = gst_asf_demux_get_uint32 (p_data, p_size); esp.buffer_size2 = gst_asf_demux_get_uint32 (p_data, p_size); esp.intial_buf_fullness2 = gst_asf_demux_get_uint32 (p_data, p_size); esp.max_obj_size = gst_asf_demux_get_uint32 (p_data, p_size); esp.flags = gst_asf_demux_get_uint32 (p_data, p_size); esp.stream_num = gst_asf_demux_get_uint16 (p_data, p_size); esp.lang_idx = gst_asf_demux_get_uint16 (p_data, p_size); esp.avg_time_per_frame = gst_asf_demux_get_uint64 (p_data, p_size); stream_name_count = gst_asf_demux_get_uint16 (p_data, p_size); payload_ext_sys_count = gst_asf_demux_get_uint16 (p_data, p_size); GST_INFO ("start_time = %" GST_TIME_FORMAT, GST_TIME_ARGS (esp.start_time * GST_MSECOND)); GST_INFO ("end_time = %" GST_TIME_FORMAT, GST_TIME_ARGS (esp.end_time * GST_MSECOND)); GST_INFO ("flags = %08x", esp.flags); GST_INFO ("average time per frame = %" GST_TIME_FORMAT, GST_TIME_ARGS (esp.avg_time_per_frame * 100)); GST_INFO ("stream number = %u", esp.stream_num); GST_INFO ("stream language ID idx = %u (%s)", esp.lang_idx, (esp.lang_idx < demux->num_languages) ? GST_STR_NULL (demux->languages[esp.lang_idx]) : "??"); GST_INFO ("stream name count = %u", stream_name_count); /* read stream names */ for (i = 0; i < stream_name_count; ++i) { guint16 stream_lang_idx; gchar *stream_name = NULL; stream_lang_idx = gst_asf_demux_get_uint16 (p_data, p_size); if (!gst_asf_demux_get_string (&stream_name, NULL, p_data, p_size)) return GST_FLOW_ERROR; GST_INFO ("stream name %d: %s", i, GST_STR_NULL (stream_name)); g_free (stream_name); /* TODO: store names in struct */ } /* read payload extension systems stuff */ GST_LOG ("payload ext sys count = %u", payload_ext_sys_count); for (i = 0; i < payload_ext_sys_count; ++i) { guint32 sys_info_len; if (!gst_asf_demux_skip_bytes (16 + 2, p_data, p_size) || *p_size < 4) return GST_FLOW_ERROR; sys_info_len = gst_asf_demux_get_uint32 (p_data, p_size); GST_LOG ("payload systems info len = %u", sys_info_len); if (!gst_asf_demux_skip_bytes (sys_info_len, p_data, p_size)) return GST_FLOW_ERROR; } GST_LOG ("bytes read: %u/%u", (guint) (*p_data - data_start), obj_size); /* there might be an optional STREAM_INFO object here now */ if ((guint) (*p_data - data_start) == obj_size) goto done; /* get size of the stream object */ if (!gst_asf_demux_get_object_header (demux, &obj_id, &len, p_data, p_size) || obj_id != ASF_OBJ_STREAM || len > (10 * 1024 * 1024)) { return GST_FLOW_ERROR; } len -= ASF_DEMUX_OBJECT_HEADER_SIZE; /* process this stream object later after all the other 'normal' ones * have been processed (since the others are more important/non-hidden) */ if (!gst_asf_demux_get_bytes (&data, (guint) len, p_data, p_size)) return GST_FLOW_ERROR; esp.stream_obj_data = data; esp.stream_obj_len = len; done: demux->ext_stream_props = g_slist_append (demux->ext_stream_props, g_memdup (&esp, sizeof (esp))); return GST_FLOW_OK; } static const gchar * gst_asf_demux_push_obj (GstASFDemux * demux, guint32 obj_id) { const gchar *nick; nick = gst_asf_get_guid_nick (asf_object_guids, obj_id); if (g_str_has_prefix (nick, "ASF_OBJ_")) nick += strlen ("ASF_OBJ_"); if (demux->objpath == NULL) { demux->objpath = g_strdup (nick); } else { gchar *newpath; newpath = g_strdup_printf ("%s/%s", demux->objpath, nick); g_free (demux->objpath); demux->objpath = newpath; } return (const gchar *) demux->objpath; } static void gst_asf_demux_pop_obj (GstASFDemux * demux) { gchar *s; if ((s = g_strrstr (demux->objpath, "/"))) { *s = '\0'; } else { g_free (demux->objpath); demux->objpath = NULL; } } static void gst_asf_demux_process_queued_extended_stream_objects (GstASFDemux * demux) { GSList *l; GST_DEBUG ("parsing stream objects embedded in extended stream properties"); for (l = demux->ext_stream_props; l != NULL; l = l->next) { asf_obj_ext_stream_properties *esp; guint64 len; guint8 *data; esp = (asf_obj_ext_stream_properties *) l->data; data = esp->stream_obj_data; len = esp->stream_obj_len; if (data && gst_asf_demux_process_stream (demux, &data, &len) != GST_FLOW_OK) { GST_WARNING_OBJECT (demux, "failed to parse stream object in extended " "stream properties object for stream %u", esp->stream_num); } g_free (esp->stream_obj_data); esp->stream_obj_data = NULL; esp->stream_obj_data = 0; } } static GstFlowReturn gst_asf_demux_process_object (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { GstFlowReturn ret; guint32 obj_id; guint64 obj_size, obj_data_size; if (!gst_asf_demux_get_object_header (demux, &obj_id, &obj_size, p_data, p_size)) { demux->bytes_needed = ASF_DEMUX_OBJECT_HEADER_SIZE; return ASF_FLOW_NEED_MORE_DATA; } obj_data_size = obj_size - ASF_DEMUX_OBJECT_HEADER_SIZE; if (obj_id != ASF_OBJ_DATA && *p_size < obj_data_size) { demux->bytes_needed = obj_size; return ASF_FLOW_NEED_MORE_DATA; } gst_asf_demux_push_obj (demux, obj_id); GST_INFO ("%s, size %" G_GUINT64_FORMAT, demux->objpath, obj_size); switch (obj_id) { case ASF_OBJ_STREAM: ret = gst_asf_demux_process_stream (demux, p_data, p_size); break; case ASF_OBJ_DATA: /* process pending stream objects and create pads for those */ gst_asf_demux_process_queued_extended_stream_objects (demux); /* switch into data mode */ ret = gst_asf_demux_process_data (demux, obj_data_size, p_data, p_size); break; case ASF_OBJ_FILE: ret = gst_asf_demux_process_file (demux, p_data, p_size); break; case ASF_OBJ_HEADER: ret = gst_asf_demux_process_header (demux, p_data, p_size); break; case ASF_OBJ_COMMENT: ret = gst_asf_demux_process_comment (demux, p_data, p_size); break; case ASF_OBJ_HEAD1: ret = gst_asf_demux_process_header_ext (demux, p_data, p_size); break; case ASF_OBJ_BITRATE_PROPS: ret = gst_asf_demux_process_bitrate_props_object (demux, p_data, p_size); break; case ASF_OBJ_EXT_CONTENT_DESC: ret = gst_asf_demux_process_ext_content_desc (demux, p_data, p_size); break; case ASF_OBJ_EXTENDED_STREAM_PROPS: ret = gst_asf_demux_process_ext_stream_props (demux, obj_data_size, p_data, p_size); break; case ASF_OBJ_LANGUAGE_LIST: ret = gst_asf_demux_process_language_list (demux, p_data, p_size); break; case ASF_OBJ_CONTENT_ENCRYPTION: case ASF_OBJ_EXT_CONTENT_ENCRYPTION: case ASF_OBJ_DIGITAL_SIGNATURE_OBJECT: goto error_encrypted; case ASF_OBJ_CONCEAL_NONE: case ASF_OBJ_HEAD2: case ASF_OBJ_UNDEFINED: case ASF_OBJ_CODEC_COMMENT: case ASF_OBJ_INDEX: case ASF_OBJ_PADDING: case ASF_OBJ_BITRATE_MUTEX: case ASF_OBJ_METADATA_OBJECT: case ASF_OBJ_COMPATIBILITY: case ASF_OBJ_INDEX_PLACEHOLDER: case ASF_OBJ_INDEX_PARAMETERS: case ASF_OBJ_ADVANCED_MUTUAL_EXCLUSION: case ASF_OBJ_STREAM_PRIORITIZATION: case ASF_OBJ_SCRIPT_COMMAND: default: /* Unknown/unhandled object read. Just ignore * it, people don't like fatal errors much */ GST_INFO ("Skipping object (size %" G_GUINT64_FORMAT ") ...", obj_size); if (!gst_asf_demux_skip_bytes (obj_size - ASF_DEMUX_OBJECT_HEADER_SIZE, p_data, p_size)) ret = ASF_FLOW_NEED_MORE_DATA; else ret = GST_FLOW_OK; break; } GST_LOG ("ret = %s (%s)", gst_asf_get_flow_name (ret), demux->objpath); gst_asf_demux_pop_obj (demux); return ret; /* ERRORS */ error_encrypted: { GST_ELEMENT_ERROR (demux, STREAM, DECODE, (_("This file is encrypted and cannot be played.")), (NULL)); return GST_FLOW_ERROR; } } static void gst_asf_demux_descramble_segment (GstASFDemux * demux, asf_segment_info * segment_info, asf_stream_context * stream) { GstBuffer *scrambled_buffer; GstBuffer *descrambled_buffer; GstBuffer *sub_buffer; guint offset; guint off; guint row; guint col; guint idx; /* descrambled_buffer is initialised in the first iteration */ descrambled_buffer = NULL; scrambled_buffer = stream->payload; if (segment_info->segment_size < demux->ds_packet_size * demux->span) return; for (offset = 0; offset < segment_info->segment_size; offset += demux->ds_chunk_size) { off = offset / demux->ds_chunk_size; row = off / demux->span; col = off % demux->span; idx = row + col * demux->ds_packet_size / demux->ds_chunk_size; GST_DEBUG ("idx=%u, row=%u, col=%u, off=%u, ds_chunk_size=%u", idx, row, col, off, demux->ds_chunk_size); GST_DEBUG ("segment_info->segment_size=%u, span=%u, packet_size=%u", segment_info->segment_size, demux->span, demux->ds_packet_size); GST_DEBUG ("GST_BUFFER_SIZE (scrambled_buffer) = %u", GST_BUFFER_SIZE (scrambled_buffer)); sub_buffer = gst_buffer_create_sub (scrambled_buffer, idx * demux->ds_chunk_size, demux->ds_chunk_size); if (!offset) { descrambled_buffer = sub_buffer; } else { GstBuffer *newbuf; newbuf = gst_buffer_merge (descrambled_buffer, sub_buffer); gst_buffer_unref (sub_buffer); gst_buffer_unref (descrambled_buffer); descrambled_buffer = newbuf; } } stream->payload = descrambled_buffer; gst_buffer_unref (scrambled_buffer); } static gboolean gst_asf_demux_element_send_event (GstElement * element, GstEvent * event) { GstASFDemux *demux = GST_ASF_DEMUX (element); gint i; GST_DEBUG ("handling element event of type %s", GST_EVENT_TYPE_NAME (event)); for (i = 0; i < demux->num_streams; ++i) { gst_event_ref (event); if (gst_asf_demux_handle_src_event (demux->stream[i].pad, event)) { gst_event_unref (event); return TRUE; } } gst_event_unref (event); return FALSE; } /* takes ownership of the passed event */ static gboolean gst_asf_demux_send_event_unlocked (GstASFDemux * demux, GstEvent * event) { gboolean ret = TRUE; gint i; GST_DEBUG ("sending event of type %s to all source pads", GST_EVENT_TYPE_NAME (event)); for (i = 0; i < demux->num_streams; ++i) { gst_event_ref (event); gst_pad_push_event (demux->stream[i].pad, event); } gst_event_unref (event); return ret; } static GstFlowReturn gst_asf_demux_push_buffer (GstASFDemux * demux, asf_stream_context * stream, GstBuffer * buf) { GstFlowReturn ret; /* do we need to send a newsegment event? */ if (stream->need_newsegment) { GST_DEBUG ("sending new-segment event on pad %s: %" GST_TIME_FORMAT " - %" GST_TIME_FORMAT, GST_PAD_NAME (stream->pad), GST_TIME_ARGS (demux->segment.start), GST_TIME_ARGS (demux->segment.stop)); /* FIXME: if we need to send a newsegment event on this pad and * the buffer doesn't have a timestamp, should we just drop the buffer * and wait for one with a timestamp before sending it? */ /* FIXME: last parameter in newsegment isn't right, is it?! */ gst_pad_push_event (stream->pad, gst_event_new_new_segment (FALSE, demux->segment.rate, GST_FORMAT_TIME, demux->segment.start, demux->segment.stop, demux->segment.start)); stream->need_newsegment = FALSE; } /* need to send tags? */ if (stream->pending_tags) { GST_LOG_OBJECT (demux, "tags %" GST_PTR_FORMAT, stream->pending_tags); gst_element_found_tags_for_pad (GST_ELEMENT (demux), stream->pad, stream->pending_tags); stream->pending_tags = NULL; } /* don't set the same time stamp on multiple consecutive outgoing * video buffers, set it on the first one and set NONE on the others, * it's the decoder's job to fill the missing bits properly */ if (stream->is_video && GST_BUFFER_TIMESTAMP_IS_VALID (buf) && GST_BUFFER_TIMESTAMP (buf) == stream->last_buffer_timestamp) { GST_BUFFER_TIMESTAMP (buf) = GST_CLOCK_TIME_NONE; } /* make sure segment.last_stop is continually increasing */ if (GST_BUFFER_TIMESTAMP_IS_VALID (buf) && demux->segment.last_stop < (gint64) GST_BUFFER_TIMESTAMP (buf)) { gst_segment_set_last_stop (&demux->segment, GST_FORMAT_TIME, GST_BUFFER_TIMESTAMP (buf)); } if (GST_BUFFER_TIMESTAMP_IS_VALID (buf)) stream->last_buffer_timestamp = GST_BUFFER_TIMESTAMP (buf); gst_buffer_set_caps (buf, stream->caps); GST_INFO ("pushing buffer on pad %s, ts=%" GST_TIME_FORMAT, GST_PAD_NAME (stream->pad), GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (buf))); ret = gst_pad_push (stream->pad, buf); if (ret == GST_FLOW_NOT_LINKED) ret = GST_FLOW_OK; return ret; } static GstFlowReturn gst_asf_demux_process_chunk (GstASFDemux * demux, asf_packet_info * packet_info, asf_segment_info * segment_info, guint8 ** p_data, guint64 * p_size) { GstFlowReturn ret = GST_FLOW_OK; asf_stream_context *stream; GstBuffer *buffer; stream = gst_asf_demux_get_stream (demux, segment_info->stream_number); if (stream == NULL) { GST_WARNING ("invalid stream number %d", segment_info->stream_number); if (!gst_asf_demux_skip_bytes (segment_info->chunk_size, p_data, p_size)) ret = ASF_FLOW_NEED_MORE_DATA; goto done; } GST_DEBUG ("Processing %s chunk of size %u (frag_offset=%d)", GST_PAD_NAME (stream->pad), segment_info->chunk_size, stream->frag_offset); if (segment_info->frag_offset == 0) { /* new packet */ stream->sequence = segment_info->sequence; if (!GST_CLOCK_TIME_IS_VALID (stream->first_pts)) stream->first_pts = segment_info->frag_timestamp - demux->preroll; demux->pts = segment_info->frag_timestamp - demux->preroll - stream->first_pts; /* if (stream->is_video) { GST_DEBUG ("%s: demux->pts=%lld (frag_timestamp=%ld, preroll=%lld)", GST_PAD_NAME (stream->pad), demux->pts, segment_info->frag_timestamp, demux->preroll); } */ if (!gst_asf_demux_get_buffer (&buffer, segment_info->chunk_size, p_data, p_size)) { return ASF_FLOW_NEED_MORE_DATA; } GST_DEBUG ("BUFFER: Copied stream to buffer %p", buffer); stream->payload = buffer; } else { GST_DEBUG ("segment_info->sequence=%d, stream->sequence=%d," " segment_info->frag_offset=%d, stream->frag_offset=%d", segment_info->sequence, stream->sequence, segment_info->frag_offset, stream->frag_offset); if (segment_info->sequence == stream->sequence && segment_info->frag_offset == stream->frag_offset) { GstBuffer *new_buffer; /* continuing packet */ GST_INFO ("continuation packet"); if (!gst_asf_demux_get_buffer (&buffer, segment_info->chunk_size, p_data, p_size)) { return ASF_FLOW_NEED_MORE_DATA; } GST_DEBUG ("copied stream to buffer %p", buffer); new_buffer = gst_buffer_merge (stream->payload, buffer); GST_DEBUG_OBJECT (demux, "BUFFER: Merged new_buffer (%p - %d) from stream->payload (%p - %d)" " and buffer (%p - %d)", new_buffer, GST_MINI_OBJECT_REFCOUNT_VALUE (new_buffer), stream->payload, GST_MINI_OBJECT_REFCOUNT_VALUE (stream->payload), buffer, GST_MINI_OBJECT_REFCOUNT_VALUE (buffer)); gst_buffer_unref (stream->payload); gst_buffer_unref (buffer); stream->payload = new_buffer; } else { /* cannot continue current packet: free it */ if (stream->frag_offset != 0) { /* cannot create new packet */ GST_DEBUG ("BUFFER: Freeing stream->payload (%p)", stream->payload); gst_buffer_unref (stream->payload); #if 0 /* FIXME: is this right/needed? we already do that below, no? */ packet_info->size_left -= segment_info->chunk_size; #endif stream->frag_offset = 0; } demux->pts = segment_info->frag_timestamp - demux->preroll - stream->first_pts; /* if (stream->is_video) { GST_DEBUG ("%s: demux->pts=%lld (frag_timestamp=%ld, preroll=%lld)", GST_PAD_NAME (stream->pad), demux->pts, segment_info->frag_timestamp, demux->preroll); } */ goto done; #if 0 /* FIXME: where did this come from / fit in ? */ return TRUE; else { /* create new packet */ stream->sequence = segment_info->sequence; } #endif } } stream->frag_offset += segment_info->chunk_size; GST_DEBUG ("frag_offset = %d segment_size = %d ", stream->frag_offset, segment_info->segment_size); if (stream->frag_offset < segment_info->segment_size) { /* We don't have the whole packet yet */ } else { /* We have the whole packet now so we should push the packet to the src pad now. First though we should check if we need to do descrambling */ if (demux->span > 1) { gst_asf_demux_descramble_segment (demux, segment_info, stream); } if (stream->is_video) { GST_DEBUG ("%s: demux->pts=%lld=%" GST_TIME_FORMAT ", stream->last_pts=%lld=%" GST_TIME_FORMAT, GST_PAD_NAME (stream->pad), demux->pts, GST_TIME_ARGS ((GST_SECOND / 1000) * demux->pts), stream->last_pts, GST_TIME_ARGS ((GST_SECOND / 1000) * stream->last_pts)); } /* FIXME: last_pts is not a GstClockTime and not in nanoseconds, so * this is not really 100% right ... */ if (demux->pts >= stream->last_pts || !GST_CLOCK_TIME_IS_VALID (stream->last_pts)) { stream->last_pts = demux->pts; } GST_BUFFER_TIMESTAMP (stream->payload) = (GST_SECOND / 1000) * stream->last_pts; GST_DEBUG ("sending stream %d of size %d", stream->id, segment_info->chunk_size); if (!stream->fps_known) { if (!stream->cache) { stream->cache = stream->payload; } else { gdouble fps; gint64 diff; gint num, denom; /* why is all this needed anyway? (tpm) */ diff = GST_BUFFER_TIMESTAMP (stream->payload) - GST_BUFFER_TIMESTAMP (stream->cache); fps = (gdouble) GST_SECOND / diff; /* artificial cap */ if (fps >= 50.0) { num = 50; denom = 1; } else if (fps <= 5.0) { num = 5; denom = 1; } else { /* crack alert */ num = (gint) GST_SECOND; while (diff > G_MAXINT) { num = num >> 1; diff = diff >> 1; } denom = (gint) diff; } stream->fps_known = TRUE; stream->caps = gst_caps_make_writable (stream->caps); gst_caps_set_simple (stream->caps, "framerate", GST_TYPE_FRACTION, num, denom, NULL); GST_DEBUG ("set up stream with fps %d/%d", num, denom); gst_pad_use_fixed_caps (stream->pad); gst_pad_set_caps (stream->pad, stream->caps); ret = gst_asf_demux_push_buffer (demux, stream, stream->cache); stream->cache = NULL; ret = gst_asf_demux_push_buffer (demux, stream, stream->payload); stream->payload = NULL; } } else { ret = gst_asf_demux_push_buffer (demux, stream, stream->payload); stream->payload = NULL; } stream->frag_offset = 0; } done: packet_info->size_left -= segment_info->chunk_size; return ret; } static GstFlowReturn gst_asf_demux_process_segment (GstASFDemux * demux, asf_packet_info * packet_info, guint8 ** p_data, guint64 * p_size) { GstFlowReturn ret = GST_FLOW_OK; asf_segment_info segment_info; gboolean has_key_frame; guint64 start_size; guint32 replic_size; guint32 time_start; guint32 frag_size; guint32 rsize; guint8 time_delta; guint8 byte; start_size = *p_size; if (*p_size < 1) return ASF_FLOW_NEED_MORE_DATA; byte = gst_asf_demux_get_uint8 (p_data, p_size); segment_info.stream_number = byte & 0x7f; has_key_frame = ((byte & 0x80) == 0x80); /* FIXME: use this somewhere? */ GST_INFO ("processing segment for stream %u%s", segment_info.stream_number, (has_key_frame) ? " (has keyframe)" : ""); /* FIXME: check (doesn't work) */ #if 0 { asf_stream_context *stream; stream = gst_asf_demux_get_stream (demux, segment_info.stream_number); if (stream && stream->last_pts == GST_CLOCK_TIME_NONE && stream->is_video && !has_key_frame) { g_print ("skipping segment, waiting for a key unit\n"); if (!gst_asf_demux_skip_bytes (segment_info.segment_size - 1, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; packet_info->size_left -= segment_info.segment_size; return GST_FLOW_OK; } } #endif segment_info.sequence = gst_asf_demux_get_var_length (packet_info->seqtype, p_data, p_size); segment_info.frag_offset = gst_asf_demux_get_var_length (packet_info->fragoffsettype, p_data, p_size); replic_size = gst_asf_demux_get_var_length (packet_info->replicsizetype, p_data, p_size); GST_DEBUG ("sequence=%u, frag_offset=%u, replic_size=%u", segment_info.sequence, segment_info.frag_offset, replic_size); if (replic_size > 1) { asf_replicated_data replicated_data_header; segment_info.compressed = FALSE; /* It's uncompressed with replic data */ if (!gst_asf_demux_get_replicated_data (&replicated_data_header, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; /* { GST_ELEMENT_ERROR (demux, STREAM, DEMUX, (NULL), ("The payload has replicated data but the size is less than 8")); return GST_FLOW_ERROR; } */ segment_info.frag_timestamp = replicated_data_header.frag_timestamp; segment_info.segment_size = replicated_data_header.object_size; if (replic_size > 8) { if (!gst_asf_demux_skip_bytes ((replic_size - 8), p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; } } else { if (replic_size == 1) { /* It's compressed */ segment_info.compressed = TRUE; time_delta = gst_asf_demux_get_uint8 (p_data, p_size); GST_DEBUG ("time_delta = %u", time_delta); } else { segment_info.compressed = FALSE; } time_start = segment_info.frag_offset; segment_info.frag_offset = 0; segment_info.frag_timestamp = time_start; } GST_DEBUG ("multiple = %u, compressed = %u", packet_info->multiple, segment_info.compressed); if (packet_info->multiple) { frag_size = gst_asf_demux_get_var_length (packet_info->segsizetype, p_data, p_size); } else { frag_size = packet_info->size_left - (start_size - *p_size); } rsize = start_size - *p_size; packet_info->size_left -= rsize; GST_DEBUG ("size left = %u, frag size = %u, rsize = %u", packet_info->size_left, frag_size, rsize); if (segment_info.compressed) { while (frag_size > 0) { byte = gst_asf_demux_get_uint8 (p_data, p_size); packet_info->size_left--; segment_info.chunk_size = byte; segment_info.segment_size = segment_info.chunk_size; if (segment_info.chunk_size > packet_info->size_left) { return ASF_FLOW_NEED_MORE_DATA; /* or is this an error? * GST_ELEMENT_ERROR (demux, STREAM, DEMUX, (NULL), * ("Payload chunk overruns packet size.")); * return GST_FLOW_ERROR; */ } ret = gst_asf_demux_process_chunk (demux, packet_info, &segment_info, p_data, p_size); if (ret != GST_FLOW_OK) break; if (segment_info.chunk_size < frag_size) frag_size -= segment_info.chunk_size + 1; else { /* GST_ELEMENT_ERROR (demux, STREAM, DEMUX, ("Invalid data in stream"), ("Invalid fragment size indicator in segment")); ret = GST_FLOW_ERROR; */ return ASF_FLOW_NEED_MORE_DATA; break; } } } else { segment_info.chunk_size = frag_size; ret = gst_asf_demux_process_chunk (demux, packet_info, &segment_info, p_data, p_size); } return ret; } static GstFlowReturn gst_asf_demux_handle_data (GstASFDemux * demux, guint8 ** p_data, guint64 * p_size) { asf_packet_info packet_info; gboolean correction; guint64 start_size; guint32 sequence; guint32 packet_length; guint32 rsize; guint16 duration; guint8 num_segments; guint8 segment; guint8 flags; guint8 property; start_size = *p_size; GST_INFO ("processing packet %" G_GUINT64_FORMAT "/%" G_GUINT64_FORMAT, demux->packet, demux->num_packets); if (demux->num_packets > 0 && demux->packet++ >= demux->num_packets) { GST_LOG ("reached EOS"); #if 0 have a gst_asf_demux_reset (demux) maybe ? gst_adapter_clear (demux->adapter); #endif gst_asf_demux_send_event_unlocked (demux, gst_event_new_eos ()); return GST_FLOW_UNEXPECTED; } if (*p_size < 1) { GST_WARNING ("unexpected end of data"); return ASF_FLOW_NEED_MORE_DATA; } correction = ((gst_asf_demux_get_uint8 (p_data, p_size) & 0x80) == 0x80); /* Uses error correction? */ if (correction) { asf_obj_data_correction corr_obj; GST_DEBUG ("data has error correction"); if (!gst_asf_demux_get_obj_data_correction (&corr_obj, p_data, p_size)) { GST_WARNING ("unexpected end of data"); return ASF_FLOW_NEED_MORE_DATA; } } /* Read the packet flags */ if (*p_size < (1 + 1)) { GST_WARNING ("unexpected end of data"); return ASF_FLOW_NEED_MORE_DATA; } flags = gst_asf_demux_get_uint8 (p_data, p_size); property = gst_asf_demux_get_uint8 (p_data, p_size); packet_info.multiple = ((flags & 0x01) == 0x01); packet_length = gst_asf_demux_get_var_length ((flags >> 5) & 0x03, p_data, p_size); sequence = gst_asf_demux_get_var_length ((flags >> 1) & 0x03, p_data, p_size); packet_info.padsize = gst_asf_demux_get_var_length ((flags >> 3) & 0x03, p_data, p_size); if (packet_length == 0) packet_length = demux->packet_size; GST_DEBUG ("multiple = %u, sequence = %u, padsize = %u, " "packet length = %u", packet_info.multiple, sequence, packet_info.padsize, packet_length); /* Read the property flags */ packet_info.replicsizetype = property & 0x03; packet_info.fragoffsettype = (property >> 2) & 0x03; packet_info.seqtype = (property >> 4) & 0x03; if (*p_size < (4 + 2)) { GST_WARNING ("unexpected end of data"); return ASF_FLOW_NEED_MORE_DATA; } demux->timestamp = gst_asf_demux_get_uint32 (p_data, p_size); duration = gst_asf_demux_get_uint16 (p_data, p_size); GST_DEBUG ("timestamp = %" GST_TIME_FORMAT ", duration = %" GST_TIME_FORMAT, GST_TIME_ARGS ((gint64) demux->timestamp * GST_MSECOND), GST_TIME_ARGS ((gint64) duration * GST_MSECOND)); /* Are there multiple payloads? */ if (packet_info.multiple) { guint8 multi_flags = gst_asf_demux_get_uint8 (p_data, p_size); packet_info.segsizetype = (multi_flags >> 6) & 0x03; num_segments = multi_flags & 0x3f; } else { packet_info.segsizetype = 2; num_segments = 1; } rsize = start_size - *p_size; packet_info.size_left = packet_length - packet_info.padsize - rsize; GST_DEBUG ("rsize: %u, size left: %u", rsize, packet_info.size_left); for (segment = 0; segment < num_segments; ++segment) { GstFlowReturn ret; ret = gst_asf_demux_process_segment (demux, &packet_info, p_data, p_size); if (ret != GST_FLOW_OK) { GST_DEBUG ("process_segment %u returned %s", segment, gst_asf_get_flow_name (ret)); return ret; } } /* Skip the padding */ if (packet_info.padsize > 0) { if (*p_size < packet_info.padsize) { GST_WARNING ("unexpected end of data"); return ASF_FLOW_NEED_MORE_DATA; } if (!gst_asf_demux_skip_bytes (packet_info.padsize, p_data, p_size)) return ASF_FLOW_NEED_MORE_DATA; } GST_DEBUG ("remaining size left: %u", packet_info.size_left); /* FIXME: this doesn't really make sense, does it? if we don't have enough * bytes left to skip the stuff at the end and we've already sent out * buffers, just returning NEED_MORE_DATA isn't really right. Should we * just throw an error in that case (can it happen with a non-broken * stream?) */ if (packet_info.size_left > 0) { if (!gst_asf_demux_skip_bytes (packet_info.size_left, p_data, p_size)) { GST_WARNING ("unexpected end of data, *p_size=%lld,packet_info.size_left=%u", *p_size, packet_info.size_left); return ASF_FLOW_NEED_MORE_DATA; } } return GST_FLOW_OK; } static GstFlowReturn gst_asf_demux_parse_data (GstASFDemux * demux) { GstFlowReturn ret = GST_FLOW_OK; /* this is basically an infinite loop */ switch (demux->state) { case GST_ASF_DEMUX_STATE_HEADER:{ guint64 data_left; guint8 *data; data_left = (guint64) gst_adapter_available (demux->adapter); GST_DEBUG ("STATE_HEADER, avail=%u:", data_left); data = (guint8 *) gst_adapter_peek (demux->adapter, data_left); ret = gst_asf_demux_process_object (demux, &data, &data_left); if (ret != ASF_FLOW_NEED_MORE_DATA) { guint bytes_used = gst_adapter_available (demux->adapter) - data_left; GST_DEBUG ("flushing %u bytes", bytes_used); gst_adapter_flush (demux->adapter, bytes_used); } else { GST_DEBUG ("not flushing, process_object returned %s", gst_asf_get_flow_name (ret)); } break; } case GST_ASF_DEMUX_STATE_DATA:{ guint64 data_size, start_data_size; guint8 *data; guint avail; avail = gst_adapter_available (demux->adapter); GST_DEBUG ("STATE_DATA, avail=%u:", avail); /* make sure a full packet is actually available */ if (demux->packet_size != (guint32) - 1 && avail < demux->packet_size) { demux->bytes_needed = demux->packet_size; return ASF_FLOW_NEED_MORE_DATA; } if (demux->packet_size == (guint32) - 1) data_size = avail; else data_size = demux->packet_size; start_data_size = data_size; data = (guint8 *) gst_adapter_peek (demux->adapter, data_size); ret = gst_asf_demux_handle_data (demux, &data, &data_size); if (ret != ASF_FLOW_NEED_MORE_DATA) { if (demux->packet_size == (guint32) - 1) { guint bytes_used = start_data_size - data_size; GST_DEBUG ("flushing %u bytes", bytes_used); gst_adapter_flush (demux->adapter, bytes_used); } else { GST_DEBUG ("flushing %u bytes", demux->packet_size); gst_adapter_flush (demux->adapter, demux->packet_size); } } else { GST_DEBUG ("not flushing, handle_data returned %s", gst_asf_get_flow_name (ret)); /* if we know the packet size and still do a * short read, then something is fishy */ if (demux->packet_size != (guint32) - 1) { /* GST_ELEMENT_ERROR (demux, STREAM, DEMUX, ("Error parsing packet"), ("Unexpected short read in packet at offset %" G_GINT64_FORMAT, gst_asf_demux_get_current_offset (demux, NULL))); ret = GST_FLOW_ERROR; */ gst_adapter_flush (demux->adapter, demux->packet_size); ret = GST_FLOW_OK; } } break; } case GST_ASF_DEMUX_STATE_EOS:{ GST_DEBUG ("STATE_EOS:"); gst_pad_event_default (demux->sinkpad, gst_event_new_eos ()); break; } default: g_return_val_if_reached (GST_FLOW_UNEXPECTED); } return ret; } static const GstQueryType * gst_asf_demux_get_src_query_types (GstPad * pad) { static const GstQueryType types[] = { GST_QUERY_POSITION, GST_QUERY_DURATION, 0 }; return types; } static gboolean gst_asf_demux_handle_src_query (GstPad * pad, GstQuery * query) { GstASFDemux *demux; gboolean res = FALSE; demux = GST_ASF_DEMUX (gst_pad_get_parent (pad)); GST_DEBUG ("handling %s query", gst_query_type_get_name (GST_QUERY_TYPE (query))); switch (GST_QUERY_TYPE (query)) { case GST_QUERY_DURATION: { GstFormat format; gst_query_parse_duration (query, &format, NULL); if (format != GST_FORMAT_TIME) { GST_LOG ("only support duration queries in TIME format"); break; } GST_OBJECT_LOCK (demux); if (demux->segment.duration != GST_CLOCK_TIME_NONE) { GST_LOG ("returning duration: %" GST_TIME_FORMAT, GST_TIME_ARGS (demux->segment.duration)); gst_query_set_duration (query, GST_FORMAT_TIME, demux->segment.duration); res = TRUE; } else { GST_LOG ("duration not known yet"); } GST_OBJECT_UNLOCK (demux); break; } case GST_QUERY_POSITION:{ GstFormat format; gst_query_parse_position (query, &format, NULL); if (format != GST_FORMAT_TIME) { GST_LOG ("only support position queries in TIME format"); break; } GST_OBJECT_LOCK (demux); if (demux->segment.last_stop != GST_CLOCK_TIME_NONE) { GST_LOG ("returning position: %" GST_TIME_FORMAT, GST_TIME_ARGS (demux->segment.last_stop)); gst_query_set_position (query, GST_FORMAT_TIME, demux->segment.last_stop); res = TRUE; } else { GST_LOG ("position not known yet"); } GST_OBJECT_UNLOCK (demux); break; } default: res = gst_pad_query_default (pad, query); break; } gst_object_unref (demux); return res; } static GstStateChangeReturn gst_asf_demux_change_state (GstElement * element, GstStateChange transition) { GstASFDemux *demux = GST_ASF_DEMUX (element); GstStateChangeReturn ret = GST_STATE_CHANGE_SUCCESS; switch (transition) { case GST_STATE_CHANGE_READY_TO_PAUSED:{ gst_segment_init (&demux->segment, GST_FORMAT_TIME); demux->adapter = gst_adapter_new (); demux->next_byte_offset = GST_BUFFER_OFFSET_NONE; break; } default: break; } ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); if (ret == GST_STATE_CHANGE_FAILURE) return ret; switch (transition) { case GST_STATE_CHANGE_PAUSED_TO_READY:{ gst_segment_init (&demux->segment, GST_FORMAT_UNDEFINED); gst_adapter_clear (demux->adapter); g_object_unref (demux->adapter); demux->adapter = NULL; if (demux->taglist) { gst_tag_list_free (demux->taglist); demux->taglist = NULL; } demux->state = GST_ASF_DEMUX_STATE_HEADER; g_free (demux->objpath); demux->objpath = NULL; g_strfreev (demux->languages); demux->languages = NULL; demux->num_languages = 0; g_slist_foreach (demux->ext_stream_props, (GFunc) g_free, NULL); g_slist_free (demux->ext_stream_props); demux->ext_stream_props = NULL; memset (demux->stream, 0, sizeof (demux->stream)); break; } default: break; } return ret; }