/* GStreamer * Copyright (C) <1999> Erik Walthinsen * Copyright (C) <2003> David A. Schleef * Copyright (C) <2006> Wim Taymans * Copyright (C) <2007> Julien Moutte * Copyright (C) <2009> Tim-Philipp Müller * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ /** * SECTION:element-qtdemux * * Demuxes a .mov file into raw or compressed audio and/or video streams. * * This element supports both push and pull-based scheduling, depending on the * capabilities of the upstream elements. * * * Example launch line * |[ * gst-launch filesrc location=test.mov ! qtdemux name=demux demux.audio_00 ! decodebin ! audioconvert ! audioresample ! autoaudiosink demux.video_00 ! queue ! decodebin ! ffmpegcolorspace ! videoscale ! autovideosink * ]| Play (parse and decode) a .mov file and try to output it to * an automatically detected soundcard and videosink. If the MOV file contains * compressed audio or video data, this will only work if you have the * right decoder elements/plugins installed. * * * Last reviewed on 2006-12-29 (0.10.5) */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "gst/gst-i18n-plugin.h" #include #include "qtatomparser.h" #include "qtdemux_types.h" #include "qtdemux_dump.h" #include "qtdemux_fourcc.h" #include "qtdemux.h" #include "qtpalette.h" #include #include #include #ifdef HAVE_ZLIB # include #endif /* max. size considered 'sane' for non-mdat atoms */ #define QTDEMUX_MAX_ATOM_SIZE (25*1024*1024) /* if the sample index is larger than this, something is likely wrong */ #define QTDEMUX_MAX_SAMPLE_INDEX_SIZE (50*1024*1024) GST_DEBUG_CATEGORY (qtdemux_debug); /*typedef struct _QtNode QtNode; */ typedef struct _QtDemuxSegment QtDemuxSegment; typedef struct _QtDemuxSample QtDemuxSample; /*struct _QtNode { guint32 type; guint8 *data; gint len; };*/ struct _QtDemuxSample { guint32 size; guint64 offset; GstClockTimeDiff pts_offset; /* Add this value to timestamp to get the pts */ guint64 timestamp; /* In GstClockTime */ guint64 duration; /* in GstClockTime */ gboolean keyframe; /* TRUE when this packet is a keyframe */ }; /* * Quicktime has tracks and segments. A track is a continuous piece of * multimedia content. The track is not always played from start to finish but * instead, pieces of the track are 'cut out' and played in sequence. This is * what the segments do. * * Inside the track we have keyframes (K) and delta frames. The track has its * own timing, which starts from 0 and extends to end. The position in the track * is called the media_time. * * The segments now describe the pieces that should be played from this track * and are basically tupples of media_time/duration/rate entries. We can have * multiple segments and they are all played after one another. An example: * * segment 1: media_time: 1 second, duration: 1 second, rate 1 * segment 2: media_time: 3 second, duration: 2 second, rate 2 * * To correctly play back this track, one must play: 1 second of media starting * from media_time 1 followed by 2 seconds of media starting from media_time 3 * at a rate of 2. * * Each of the segments will be played at a specific time, the first segment at * time 0, the second one after the duration of the first one, etc.. Note that * the time in resulting playback is not identical to the media_time of the * track anymore. * * Visually, assuming the track has 4 second of media_time: * * (a) (b) (c) (d) * .-----------------------------------------------------------. * track: | K.....K.........K........K.......K.......K...........K... | * '-----------------------------------------------------------' * 0 1 2 3 4 * .------------^ ^ .----------^ ^ * / .-------------' / .------------------' * / / .-----' / * .--------------. .--------------. * | segment 1 | | segment 2 | * '--------------' '--------------' * * The challenge here is to cut out the right pieces of the track for each of * the playback segments. This fortunatly can easily be done with the SEGMENT * events of gstreamer. * * For playback of segment 1, we need to provide the decoder with the keyframe * (a), in the above figure, but we must instruct it only to output the decoded * data between second 1 and 2. We do this with a SEGMENT event for 1 to 2, time * position set to the time of the segment: 0. * * We then proceed to push data from keyframe (a) to frame (b). The decoder * decodes but clips all before media_time 1. * * After finishing a segment, we push out a new SEGMENT event with the clipping * boundaries of the new data. * * This is a good usecase for the GStreamer accumulated SEGMENT events. */ struct _QtDemuxSegment { /* global time and duration, all gst time */ guint64 time; guint64 stop_time; guint64 duration; /* media time of trak, all gst time */ guint64 media_start; guint64 media_stop; gdouble rate; }; struct _QtDemuxStream { GstPad *pad; /* stream type */ guint32 subtype; GstCaps *caps; guint32 fourcc; /* duration/scale */ guint64 duration; /* in timescale */ guint32 timescale; /* our samples */ guint32 n_samples; QtDemuxSample *samples; gboolean all_keyframe; /* TRUE when all samples are keyframes (no stss) */ guint32 min_duration; /* duration in timescale of first sample, used for figuring out the framerate, in timescale units */ /* if we use chunks or samples */ gboolean sampled; guint padding; /* video info */ gint width; gint height; /* aspect ratio */ gint display_width; gint display_height; gint par_w; gint par_h; /* Numerator/denominator framerate */ gint fps_n; gint fps_d; guint16 bits_per_sample; guint16 color_table_id; /* audio info */ gdouble rate; gint n_channels; guint samples_per_packet; guint samples_per_frame; guint bytes_per_packet; guint bytes_per_sample; guint bytes_per_frame; guint compression; /* when a discontinuity is pending */ gboolean discont; /* list of buffers to push first */ GSList *buffers; /* if we need to clip this buffer. This is only needed for uncompressed * data */ gboolean need_clip; /* buffer needs some custom processing, e.g. subtitles */ gboolean need_process; /* current position */ guint32 segment_index; guint32 sample_index; guint64 time_position; /* in gst time */ /* the Gst segment we are processing out, used for clipping */ GstSegment segment; /* last GstFlowReturn */ GstFlowReturn last_ret; /* quicktime segments */ guint32 n_segments; QtDemuxSegment *segments; guint32 from_sample; guint32 to_sample; gboolean sent_eos; GstTagList *pending_tags; gboolean send_global_tags; GstEvent *pending_event; }; enum QtDemuxState { QTDEMUX_STATE_INITIAL, /* Initial state (haven't got the header yet) */ QTDEMUX_STATE_HEADER, /* Parsing the header */ QTDEMUX_STATE_MOVIE, /* Parsing/Playing the media data */ QTDEMUX_STATE_BUFFER_MDAT /* Buffering the mdat atom */ }; static GNode *qtdemux_tree_get_child_by_type (GNode * node, guint32 fourcc); static GNode *qtdemux_tree_get_child_by_type_full (GNode * node, guint32 fourcc, QtAtomParser * parser); static GNode *qtdemux_tree_get_sibling_by_type (GNode * node, guint32 fourcc); static const GstElementDetails gst_qtdemux_details = GST_ELEMENT_DETAILS ("QuickTime demuxer", "Codec/Demuxer", "Demultiplex a QuickTime file into audio and video streams", "David Schleef , Wim Taymans "); static GstStaticPadTemplate gst_qtdemux_sink_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS ("video/quicktime; video/mj2; audio/x-m4a; " "application/x-3gp") ); static GstStaticPadTemplate gst_qtdemux_videosrc_template = GST_STATIC_PAD_TEMPLATE ("video_%02d", GST_PAD_SRC, GST_PAD_SOMETIMES, GST_STATIC_CAPS_ANY); static GstStaticPadTemplate gst_qtdemux_audiosrc_template = GST_STATIC_PAD_TEMPLATE ("audio_%02d", GST_PAD_SRC, GST_PAD_SOMETIMES, GST_STATIC_CAPS_ANY); static GstStaticPadTemplate gst_qtdemux_subsrc_template = GST_STATIC_PAD_TEMPLATE ("subtitle_%02d", GST_PAD_SRC, GST_PAD_SOMETIMES, GST_STATIC_CAPS_ANY); static GstElementClass *parent_class = NULL; static void gst_qtdemux_class_init (GstQTDemuxClass * klass); static void gst_qtdemux_base_init (GstQTDemuxClass * klass); static void gst_qtdemux_init (GstQTDemux * quicktime_demux); static void gst_qtdemux_dispose (GObject * object); static GstStateChangeReturn gst_qtdemux_change_state (GstElement * element, GstStateChange transition); static gboolean qtdemux_sink_activate (GstPad * sinkpad); static gboolean qtdemux_sink_activate_pull (GstPad * sinkpad, gboolean active); static gboolean qtdemux_sink_activate_push (GstPad * sinkpad, gboolean active); static void gst_qtdemux_loop (GstPad * pad); static GstFlowReturn gst_qtdemux_chain (GstPad * sinkpad, GstBuffer * inbuf); static gboolean gst_qtdemux_handle_sink_event (GstPad * pad, GstEvent * event); static gboolean qtdemux_parse_moov (GstQTDemux * qtdemux, const guint8 * buffer, guint length); static gboolean qtdemux_parse_node (GstQTDemux * qtdemux, GNode * node, const guint8 * buffer, guint length); static gboolean qtdemux_parse_tree (GstQTDemux * qtdemux); static void gst_qtdemux_handle_esds (GstQTDemux * qtdemux, QtDemuxStream * stream, GNode * esds, GstTagList * list); static GstCaps *qtdemux_video_caps (GstQTDemux * qtdemux, QtDemuxStream * stream, guint32 fourcc, const guint8 * stsd_data, gchar ** codec_name); static GstCaps *qtdemux_audio_caps (GstQTDemux * qtdemux, QtDemuxStream * stream, guint32 fourcc, const guint8 * data, int len, gchar ** codec_name); static GstCaps *qtdemux_sub_caps (GstQTDemux * qtdemux, QtDemuxStream * stream, guint32 fourcc, const guint8 * data, gchar ** codec_name); GType gst_qtdemux_get_type (void) { static GType qtdemux_type = 0; if (G_UNLIKELY (!qtdemux_type)) { static const GTypeInfo qtdemux_info = { sizeof (GstQTDemuxClass), (GBaseInitFunc) gst_qtdemux_base_init, NULL, (GClassInitFunc) gst_qtdemux_class_init, NULL, NULL, sizeof (GstQTDemux), 0, (GInstanceInitFunc) gst_qtdemux_init, }; qtdemux_type = g_type_register_static (GST_TYPE_ELEMENT, "GstQTDemux", &qtdemux_info, 0); } return qtdemux_type; } static void gst_qtdemux_base_init (GstQTDemuxClass * klass) { GstElementClass *element_class = GST_ELEMENT_CLASS (klass); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_qtdemux_sink_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_qtdemux_videosrc_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_qtdemux_audiosrc_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_qtdemux_subsrc_template)); gst_element_class_set_details (element_class, &gst_qtdemux_details); GST_DEBUG_CATEGORY_INIT (qtdemux_debug, "qtdemux", 0, "qtdemux plugin"); } static void gst_qtdemux_class_init (GstQTDemuxClass * klass) { GObjectClass *gobject_class; GstElementClass *gstelement_class; gobject_class = (GObjectClass *) klass; gstelement_class = (GstElementClass *) klass; parent_class = g_type_class_peek_parent (klass); gobject_class->dispose = gst_qtdemux_dispose; gstelement_class->change_state = gst_qtdemux_change_state; } static void gst_qtdemux_init (GstQTDemux * qtdemux) { qtdemux->sinkpad = gst_pad_new_from_static_template (&gst_qtdemux_sink_template, "sink"); gst_pad_set_activate_function (qtdemux->sinkpad, qtdemux_sink_activate); gst_pad_set_activatepull_function (qtdemux->sinkpad, qtdemux_sink_activate_pull); gst_pad_set_activatepush_function (qtdemux->sinkpad, qtdemux_sink_activate_push); gst_pad_set_chain_function (qtdemux->sinkpad, gst_qtdemux_chain); gst_pad_set_event_function (qtdemux->sinkpad, gst_qtdemux_handle_sink_event); gst_element_add_pad (GST_ELEMENT_CAST (qtdemux), qtdemux->sinkpad); qtdemux->state = QTDEMUX_STATE_INITIAL; qtdemux->pullbased = FALSE; qtdemux->neededbytes = 16; qtdemux->todrop = 0; qtdemux->adapter = gst_adapter_new (); qtdemux->offset = 0; qtdemux->mdatoffset = GST_CLOCK_TIME_NONE; qtdemux->mdatbuffer = NULL; gst_segment_init (&qtdemux->segment, GST_FORMAT_TIME); } static void gst_qtdemux_dispose (GObject * object) { GstQTDemux *qtdemux = GST_QTDEMUX (object); if (qtdemux->adapter) { g_object_unref (G_OBJECT (qtdemux->adapter)); qtdemux->adapter = NULL; } G_OBJECT_CLASS (parent_class)->dispose (object); } static GstFlowReturn gst_qtdemux_pull_atom (GstQTDemux * qtdemux, guint64 offset, guint64 size, GstBuffer ** buf) { GstFlowReturn flow; /* Sanity check: catch bogus sizes (fuzzed/broken files) */ if (G_UNLIKELY (size > QTDEMUX_MAX_ATOM_SIZE)) { GST_ELEMENT_ERROR (qtdemux, STREAM, DECODE, (_("This file is invalid and cannot be played.")), ("atom has bogus size %" G_GUINT64_FORMAT, size)); return GST_FLOW_ERROR; } flow = gst_pad_pull_range (qtdemux->sinkpad, offset, size, buf); if (G_UNLIKELY (flow != GST_FLOW_OK)) return flow; /* Catch short reads - we don't want any partial atoms */ if (G_UNLIKELY (GST_BUFFER_SIZE (*buf) < size)) { GST_WARNING_OBJECT (qtdemux, "short read: %u < %" G_GUINT64_FORMAT, GST_BUFFER_SIZE (*buf), size); gst_buffer_unref (*buf); *buf = NULL; return GST_FLOW_UNEXPECTED; } return flow; } #if 0 static gboolean gst_qtdemux_src_convert (GstPad * pad, GstFormat src_format, gint64 src_value, GstFormat * dest_format, gint64 * dest_value) { gboolean res = TRUE; QtDemuxStream *stream = gst_pad_get_element_private (pad); if (stream->subtype == GST_MAKE_FOURCC ('v', 'i', 'd', 'e') && (src_format == GST_FORMAT_BYTES || *dest_format == GST_FORMAT_BYTES)) return FALSE; switch (src_format) { case GST_FORMAT_TIME: switch (*dest_format) { case GST_FORMAT_BYTES: *dest_value = src_value * 1; /* FIXME */ break; case GST_FORMAT_DEFAULT: *dest_value = src_value * 1; /* FIXME */ break; default: res = FALSE; break; } break; case GST_FORMAT_BYTES: switch (*dest_format) { case GST_FORMAT_TIME: *dest_value = src_value * 1; /* FIXME */ break; default: res = FALSE; break; } break; case GST_FORMAT_DEFAULT: switch (*dest_format) { case GST_FORMAT_TIME: *dest_value = src_value * 1; /* FIXME */ break; default: res = FALSE; break; } break; default: res = FALSE; } return res; } #endif static const GstQueryType * gst_qtdemux_get_src_query_types (GstPad * pad) { static const GstQueryType src_types[] = { GST_QUERY_POSITION, GST_QUERY_DURATION, GST_QUERY_SEEKING, 0 }; return src_types; } static gboolean gst_qtdemux_get_duration (GstQTDemux * qtdemux, gint64 * duration) { gboolean res = TRUE; *duration = GST_CLOCK_TIME_NONE; if (qtdemux->duration != 0) { if (qtdemux->duration != G_MAXINT32 && qtdemux->timescale != 0) { *duration = gst_util_uint64_scale (qtdemux->duration, GST_SECOND, qtdemux->timescale); } } return res; } static gboolean gst_qtdemux_handle_src_query (GstPad * pad, GstQuery * query) { gboolean res = FALSE; GstQTDemux *qtdemux = GST_QTDEMUX (gst_pad_get_parent (pad)); GST_LOG_OBJECT (pad, "%s query", GST_QUERY_TYPE_NAME (query)); switch (GST_QUERY_TYPE (query)) { case GST_QUERY_POSITION: if (GST_CLOCK_TIME_IS_VALID (qtdemux->segment.last_stop)) { gst_query_set_position (query, GST_FORMAT_TIME, qtdemux->segment.last_stop); res = TRUE; } break; case GST_QUERY_DURATION:{ GstFormat fmt; gst_query_parse_duration (query, &fmt, NULL); if (fmt == GST_FORMAT_TIME) { gint64 duration = -1; gst_qtdemux_get_duration (qtdemux, &duration); if (duration > 0) { gst_query_set_duration (query, GST_FORMAT_TIME, duration); res = TRUE; } } break; } case GST_QUERY_SEEKING:{ GstFormat fmt; gboolean seekable; gst_query_parse_seeking (query, &fmt, NULL, NULL, NULL); if (fmt == GST_FORMAT_TIME) { gint64 duration = -1; gst_qtdemux_get_duration (qtdemux, &duration); seekable = TRUE; if (!qtdemux->pullbased) { GstQuery *q; /* we might be able with help from upstream */ seekable = FALSE; q = gst_query_new_seeking (GST_FORMAT_BYTES); if (gst_pad_peer_query (qtdemux->sinkpad, q)) { gst_query_parse_seeking (q, &fmt, &seekable, NULL, NULL); GST_LOG_OBJECT (qtdemux, "upstream BYTE seekable %d", seekable); } gst_query_unref (q); } gst_query_set_seeking (query, GST_FORMAT_TIME, seekable, 0, duration); res = TRUE; } break; } default: res = gst_pad_query_default (pad, query); break; } gst_object_unref (qtdemux); return res; } static void gst_qtdemux_push_tags (GstQTDemux * qtdemux, QtDemuxStream * stream) { if (G_LIKELY (stream->pad)) { GST_DEBUG_OBJECT (qtdemux, "Checking pad %s:%s for tags", GST_DEBUG_PAD_NAME (stream->pad)); if (G_UNLIKELY (stream->pending_tags)) { GST_DEBUG_OBJECT (qtdemux, "Sending tags %" GST_PTR_FORMAT, stream->pending_tags); gst_pad_push_event (stream->pad, gst_event_new_tag (stream->pending_tags)); stream->pending_tags = NULL; } if (G_UNLIKELY (stream->send_global_tags && qtdemux->tag_list)) { GST_DEBUG_OBJECT (qtdemux, "Sending global tags %" GST_PTR_FORMAT, qtdemux->tag_list); gst_pad_push_event (stream->pad, gst_event_new_tag (gst_tag_list_copy (qtdemux->tag_list))); stream->send_global_tags = FALSE; } } } /* push event on all source pads; takes ownership of the event */ static void gst_qtdemux_push_event (GstQTDemux * qtdemux, GstEvent * event) { guint n; GST_DEBUG_OBJECT (qtdemux, "pushing %s event on all source pads", GST_EVENT_TYPE_NAME (event)); for (n = 0; n < qtdemux->n_streams; n++) { GstPad *pad; if ((pad = qtdemux->streams[n]->pad)) gst_pad_push_event (pad, gst_event_ref (event)); } gst_event_unref (event); } /* push a pending newsegment event, if any from the streaming thread */ static void gst_qtdemux_push_pending_newsegment (GstQTDemux * qtdemux) { if (qtdemux->pending_newsegment) { gst_qtdemux_push_event (qtdemux, qtdemux->pending_newsegment); qtdemux->pending_newsegment = NULL; } } typedef struct { guint64 media_time; } FindData; static gint find_func (QtDemuxSample * s1, guint64 * media_time, gpointer user_data) { if (s1->timestamp > *media_time) return 1; return -1; } /* find the index of the sample that includes the data for @media_time * * Returns the index of the sample. */ static guint32 gst_qtdemux_find_index (GstQTDemux * qtdemux, QtDemuxStream * str, guint64 media_time) { QtDemuxSample *result; guint32 index; result = gst_util_array_binary_search (str->samples, str->n_samples, sizeof (QtDemuxSample), (GCompareDataFunc) find_func, GST_SEARCH_MODE_BEFORE, &media_time, NULL); if (G_LIKELY (result)) index = result - str->samples; else index = 0; return index; } /* find the index of the keyframe needed to decode the sample at @index * of stream @str. * * Returns the index of the keyframe. */ static guint32 gst_qtdemux_find_keyframe (GstQTDemux * qtdemux, QtDemuxStream * str, guint32 index) { guint32 new_index = index; if (index >= str->n_samples) { new_index = str->n_samples; goto beach; } /* all keyframes, return index */ if (str->all_keyframe) { new_index = index; goto beach; } /* else go back until we have a keyframe */ while (TRUE) { if (str->samples[new_index].keyframe) break; if (new_index == 0) break; new_index--; } beach: GST_DEBUG_OBJECT (qtdemux, "searching for keyframe index before index %u " "gave %u", index, new_index); return new_index; } /* find the segment for @time_position for @stream * * Returns -1 if the segment cannot be found. */ static guint32 gst_qtdemux_find_segment (GstQTDemux * qtdemux, QtDemuxStream * stream, guint64 time_position) { gint i; guint32 seg_idx; GST_LOG_OBJECT (qtdemux, "finding segment for %" GST_TIME_FORMAT, GST_TIME_ARGS (time_position)); /* find segment corresponding to time_position if we are looking * for a segment. */ seg_idx = -1; for (i = 0; i < stream->n_segments; i++) { QtDemuxSegment *segment = &stream->segments[i]; GST_LOG_OBJECT (qtdemux, "looking at segment %" GST_TIME_FORMAT "-%" GST_TIME_FORMAT, GST_TIME_ARGS (segment->time), GST_TIME_ARGS (segment->stop_time)); /* For the last segment we include stop_time in the last segment */ if (i < stream->n_segments - 1) { if (segment->time <= time_position && time_position < segment->stop_time) { GST_LOG_OBJECT (qtdemux, "segment %d matches", i); seg_idx = i; break; } } else { if (segment->time <= time_position && time_position <= segment->stop_time) { GST_LOG_OBJECT (qtdemux, "segment %d matches", i); seg_idx = i; break; } } } return seg_idx; } /* move the stream @str to the sample position @index. * * Updates @str->sample_index and marks discontinuity if needed. */ static void gst_qtdemux_move_stream (GstQTDemux * qtdemux, QtDemuxStream * str, guint32 index) { /* no change needed */ if (index == str->sample_index) return; GST_DEBUG_OBJECT (qtdemux, "moving to sample %u of %u", index, str->n_samples); /* position changed, we have a discont */ str->sample_index = index; /* Each time we move in the stream we store the position where we are * starting from */ str->from_sample = index; str->discont = TRUE; } static void gst_qtdemux_adjust_seek (GstQTDemux * qtdemux, gint64 desired_time, gint64 * key_time, gint64 * key_offset) { guint64 min_offset; gint64 min_byte_offset = -1; gint n; min_offset = desired_time; /* for each stream, find the index of the sample in the segment * and move back to the previous keyframe. */ for (n = 0; n < qtdemux->n_streams; n++) { QtDemuxStream *str; guint32 index, kindex; guint32 seg_idx; guint64 media_start; guint64 media_time; guint64 seg_time; QtDemuxSegment *seg; str = qtdemux->streams[n]; seg_idx = gst_qtdemux_find_segment (qtdemux, str, desired_time); GST_DEBUG_OBJECT (qtdemux, "align segment %d", seg_idx); /* segment not found, continue with normal flow */ if (seg_idx == -1) continue; /* get segment and time in the segment */ seg = &str->segments[seg_idx]; seg_time = desired_time - seg->time; /* get the media time in the segment */ media_start = seg->media_start + seg_time; /* get the index of the sample with media time */ index = gst_qtdemux_find_index (qtdemux, str, media_start); GST_DEBUG_OBJECT (qtdemux, "sample for %" GST_TIME_FORMAT " at %u", GST_TIME_ARGS (media_start), index); /* find previous keyframe */ kindex = gst_qtdemux_find_keyframe (qtdemux, str, index); /* if the keyframe is at a different position, we need to update the * requested seek time */ if (index != kindex) { index = kindex; /* get timestamp of keyframe */ media_time = str->samples[kindex].timestamp; GST_DEBUG_OBJECT (qtdemux, "keyframe at %u with time %" GST_TIME_FORMAT, kindex, GST_TIME_ARGS (media_time)); /* keyframes in the segment get a chance to change the * desired_offset. keyframes out of the segment are * ignored. */ if (media_time >= seg->media_start) { guint64 seg_time; /* this keyframe is inside the segment, convert back to * segment time */ seg_time = (media_time - seg->media_start) + seg->time; if (seg_time < min_offset) min_offset = seg_time; } } if (min_byte_offset < 0 || str->samples[index].offset < min_byte_offset) min_byte_offset = str->samples[index].offset; } if (key_time) *key_time = min_offset; if (key_offset) *key_offset = min_byte_offset; } static gboolean gst_qtdemux_convert_seek (GstPad * pad, GstFormat * format, GstSeekType cur_type, gint64 * cur, GstSeekType stop_type, gint64 * stop) { gboolean res; GstFormat fmt; g_return_val_if_fail (format != NULL, FALSE); g_return_val_if_fail (cur != NULL, FALSE); g_return_val_if_fail (stop != NULL, FALSE); if (*format == GST_FORMAT_TIME) return TRUE; fmt = GST_FORMAT_TIME; res = TRUE; if (cur_type != GST_SEEK_TYPE_NONE) res = gst_pad_query_convert (pad, *format, *cur, &fmt, cur); if (res && stop_type != GST_SEEK_TYPE_NONE) res = gst_pad_query_convert (pad, *format, *stop, &fmt, stop); if (res) *format = GST_FORMAT_TIME; return res; } /* perform seek in push based mode: find BYTE position to move to based on time and delegate to upstream */ static gboolean gst_qtdemux_do_push_seek (GstQTDemux * qtdemux, GstPad * pad, GstEvent * event) { gdouble rate; GstFormat format; GstSeekFlags flags; GstSeekType cur_type, stop_type; gint64 cur, stop; gboolean res; gint64 byte_cur; GST_DEBUG_OBJECT (qtdemux, "doing push-based seek"); gst_event_parse_seek (event, &rate, &format, &flags, &cur_type, &cur, &stop_type, &stop); if (stop_type != GST_SEEK_TYPE_NONE) goto unsupported_seek; stop = -1; /* only forward streaming and seeking is possible */ if (rate <= 0) goto unsupported_seek; /* convert to TIME if needed and possible */ if (!gst_qtdemux_convert_seek (pad, &format, cur_type, &cur, stop_type, &stop)) goto no_format; /* find reasonable corresponding BYTE position, * also try to mind about keyframes, since we can not go back a bit for them * later on */ gst_qtdemux_adjust_seek (qtdemux, cur, NULL, &byte_cur); if (byte_cur == -1) goto abort_seek; GST_DEBUG_OBJECT (qtdemux, "Pushing BYTE seek rate %g, " "start %" G_GINT64_FORMAT ", stop %" G_GINT64_FORMAT, rate, byte_cur, stop); /* BYTE seek event */ event = gst_event_new_seek (rate, GST_FORMAT_BYTES, flags, cur_type, byte_cur, stop_type, stop); res = gst_pad_push_event (qtdemux->sinkpad, event); return res; /* ERRORS */ abort_seek: { GST_DEBUG_OBJECT (qtdemux, "could not determine byte position to seek to, " "seek aborted."); return FALSE; } unsupported_seek: { GST_DEBUG_OBJECT (qtdemux, "unsupported seek, seek aborted."); return FALSE; } no_format: { GST_DEBUG_OBJECT (qtdemux, "unsupported format given, seek aborted."); return FALSE; } } /* perform the seek. * * We set all segment_indexes in the streams to unknown and * adjust the time_position to the desired position. this is enough * to trigger a segment switch in the streaming thread to start * streaming from the desired position. * * Keyframe seeking is a little more complicated when dealing with * segments. Ideally we want to move to the previous keyframe in * the segment but there might not be a keyframe in the segment. In * fact, none of the segments could contain a keyframe. We take a * practical approach: seek to the previous keyframe in the segment, * if there is none, seek to the beginning of the segment. * * Called with STREAM_LOCK */ static gboolean gst_qtdemux_perform_seek (GstQTDemux * qtdemux, GstSegment * segment) { gint64 desired_offset; gint n; desired_offset = segment->last_stop; GST_DEBUG_OBJECT (qtdemux, "seeking to %" GST_TIME_FORMAT, GST_TIME_ARGS (desired_offset)); if (segment->flags & GST_SEEK_FLAG_KEY_UNIT) { gint64 min_offset; gst_qtdemux_adjust_seek (qtdemux, desired_offset, &min_offset, NULL); GST_DEBUG_OBJECT (qtdemux, "keyframe seek, align to %" GST_TIME_FORMAT, GST_TIME_ARGS (min_offset)); desired_offset = min_offset; } /* and set all streams to the final position */ for (n = 0; n < qtdemux->n_streams; n++) { QtDemuxStream *stream = qtdemux->streams[n]; stream->time_position = desired_offset; stream->sample_index = -1; stream->segment_index = -1; stream->last_ret = GST_FLOW_OK; stream->sent_eos = FALSE; } segment->last_stop = desired_offset; segment->time = desired_offset; /* we stop at the end */ if (segment->stop == -1) segment->stop = segment->duration; return TRUE; } /* do a seek in pull based mode */ static gboolean gst_qtdemux_do_seek (GstQTDemux * qtdemux, GstPad * pad, GstEvent * event) { gdouble rate; GstFormat format; GstSeekFlags flags; GstSeekType cur_type, stop_type; gint64 cur, stop; gboolean flush; gboolean update; GstSegment seeksegment; int i; if (event) { GST_DEBUG_OBJECT (qtdemux, "doing seek with event"); gst_event_parse_seek (event, &rate, &format, &flags, &cur_type, &cur, &stop_type, &stop); /* we have to have a format as the segment format. Try to convert * if not. */ if (!gst_qtdemux_convert_seek (pad, &format, cur_type, &cur, stop_type, &stop)) goto no_format; GST_DEBUG_OBJECT (qtdemux, "seek format %s", gst_format_get_name (format)); } else { GST_DEBUG_OBJECT (qtdemux, "doing seek without event"); flags = 0; } flush = flags & GST_SEEK_FLAG_FLUSH; /* stop streaming, either by flushing or by pausing the task */ if (flush) { /* unlock upstream pull_range */ gst_pad_push_event (qtdemux->sinkpad, gst_event_new_flush_start ()); /* make sure out loop function exits */ gst_qtdemux_push_event (qtdemux, gst_event_new_flush_start ()); } else { /* non flushing seek, pause the task */ gst_pad_pause_task (qtdemux->sinkpad); } /* wait for streaming to finish */ GST_PAD_STREAM_LOCK (qtdemux->sinkpad); /* copy segment, we need this because we still need the old * segment when we close the current segment. */ memcpy (&seeksegment, &qtdemux->segment, sizeof (GstSegment)); if (event) { /* configure the segment with the seek variables */ GST_DEBUG_OBJECT (qtdemux, "configuring seek"); gst_segment_set_seek (&seeksegment, rate, format, flags, cur_type, cur, stop_type, stop, &update); } /* now do the seek, this actually never returns FALSE */ gst_qtdemux_perform_seek (qtdemux, &seeksegment); /* prepare for streaming again */ if (flush) { gst_pad_push_event (qtdemux->sinkpad, gst_event_new_flush_stop ()); gst_qtdemux_push_event (qtdemux, gst_event_new_flush_stop ()); } else if (qtdemux->segment_running) { /* we are running the current segment and doing a non-flushing seek, * close the segment first based on the last_stop. */ GST_DEBUG_OBJECT (qtdemux, "closing running segment %" G_GINT64_FORMAT " to %" G_GINT64_FORMAT, qtdemux->segment.start, qtdemux->segment.last_stop); if (qtdemux->segment.rate >= 0) { /* FIXME, rate is the product of the global rate and the (quicktime) * segment rate. */ qtdemux->pending_newsegment = gst_event_new_new_segment (TRUE, qtdemux->segment.rate, qtdemux->segment.format, qtdemux->segment.start, qtdemux->segment.last_stop, qtdemux->segment.time); } else { /* For Reverse Playback */ guint64 stop; if ((stop = qtdemux->segment.stop) == -1) stop = qtdemux->segment.duration; /* for reverse playback, we played from stop to last_stop. */ qtdemux->pending_newsegment = gst_event_new_new_segment (TRUE, qtdemux->segment.rate, qtdemux->segment.format, qtdemux->segment.last_stop, stop, qtdemux->segment.last_stop); } } /* commit the new segment */ memcpy (&qtdemux->segment, &seeksegment, sizeof (GstSegment)); if (qtdemux->segment.flags & GST_SEEK_FLAG_SEGMENT) { gst_element_post_message (GST_ELEMENT_CAST (qtdemux), gst_message_new_segment_start (GST_OBJECT_CAST (qtdemux), qtdemux->segment.format, qtdemux->segment.last_stop)); } /* restart streaming, NEWSEGMENT will be sent from the streaming * thread. */ qtdemux->segment_running = TRUE; for (i = 0; i < qtdemux->n_streams; i++) qtdemux->streams[i]->last_ret = GST_FLOW_OK; gst_pad_start_task (qtdemux->sinkpad, (GstTaskFunction) gst_qtdemux_loop, qtdemux->sinkpad); GST_PAD_STREAM_UNLOCK (qtdemux->sinkpad); return TRUE; /* ERRORS */ no_format: { GST_DEBUG_OBJECT (qtdemux, "unsupported format given, seek aborted."); return FALSE; } } static gboolean gst_qtdemux_handle_src_event (GstPad * pad, GstEvent * event) { gboolean res = TRUE; GstQTDemux *qtdemux = GST_QTDEMUX (gst_pad_get_parent (pad)); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_SEEK: if (qtdemux->pullbased) { res = gst_qtdemux_do_seek (qtdemux, pad, event); } else if (qtdemux->state == QTDEMUX_STATE_MOVIE && qtdemux->n_streams) { res = gst_qtdemux_do_push_seek (qtdemux, pad, event); } else { GST_DEBUG_OBJECT (qtdemux, "ignoring seek in push mode in current state"); res = FALSE; } gst_event_unref (event); break; case GST_EVENT_QOS: case GST_EVENT_NAVIGATION: res = FALSE; gst_event_unref (event); break; default: res = gst_pad_event_default (pad, event); break; } gst_object_unref (qtdemux); return res; } /* stream/index return sample that is min/max w.r.t. byte position, * time is min/max w.r.t. time of samples, * the latter need not be time of the former sample */ static void gst_qtdemux_find_sample (GstQTDemux * qtdemux, gint64 byte_pos, gboolean fw, gboolean set, QtDemuxStream ** _stream, gint * _index, gint64 * _time) { gint i, n, index; gint64 time, min_time; QtDemuxStream *stream; min_time = -1; stream = NULL; index = -1; for (n = 0; n < qtdemux->n_streams; ++n) { QtDemuxStream *str; gint inc; gboolean set_sample; str = qtdemux->streams[n]; set_sample = !set; if (fw) { i = 0; inc = 1; } else { i = str->n_samples - 1; inc = -1; } for (; (i >= 0) && (i < str->n_samples); i += inc) { if (str->samples[i].size && ((fw && (str->samples[i].offset >= byte_pos)) || (!fw && (str->samples[i].offset + str->samples[i].size <= byte_pos)))) { /* move stream to first available sample */ if (set) { gst_qtdemux_move_stream (qtdemux, str, i); set_sample = TRUE; } /* determine min/max time */ time = str->samples[i].timestamp + str->samples[i].pts_offset; if (min_time == -1 || (fw && min_time > time) || (!fw && min_time < time)) { min_time = time; } /* determine stream with leading sample, to get its position */ /* only needed in fw case */ if (fw && (!stream || str->samples[i].offset < stream->samples[index].offset)) { stream = str; index = i; } break; } } /* no sample for this stream, mark eos */ if (!set_sample) gst_qtdemux_move_stream (qtdemux, str, str->n_samples); } if (_time) *_time = min_time; if (_stream) *_stream = stream; if (_index) *_index = index; } static gboolean gst_qtdemux_handle_sink_event (GstPad * sinkpad, GstEvent * event) { GstQTDemux *demux = GST_QTDEMUX (GST_PAD_PARENT (sinkpad)); gboolean res; GST_LOG_OBJECT (demux, "handling %s event", GST_EVENT_TYPE_NAME (event)); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_NEWSEGMENT: { GstFormat format; gdouble rate, arate; gint64 start, stop, time, offset = 0; QtDemuxStream *stream; gint idx; gboolean update; GstSegment segment; /* some debug output */ gst_segment_init (&segment, GST_FORMAT_UNDEFINED); gst_event_parse_new_segment_full (event, &update, &rate, &arate, &format, &start, &stop, &time); gst_segment_set_newsegment_full (&segment, update, rate, arate, format, start, stop, time); GST_DEBUG_OBJECT (demux, "received format %d newsegment %" GST_SEGMENT_FORMAT, format, &segment); /* chain will send initial newsegment after pads have been added */ if (demux->state != QTDEMUX_STATE_MOVIE || !demux->n_streams) { GST_DEBUG_OBJECT (demux, "still starting, eating event"); goto exit; } /* we only expect a BYTE segment, e.g. following a seek */ if (format == GST_FORMAT_BYTES) { if (start > 0) { offset = start; gst_qtdemux_find_sample (demux, start, TRUE, FALSE, NULL, NULL, &start); start = MAX (start, 0); } if (stop > 0) { gst_qtdemux_find_sample (demux, stop, FALSE, FALSE, NULL, NULL, &stop); stop = MAX (stop, 0); } } else { GST_DEBUG_OBJECT (demux, "unsupported segment format, ignoring"); goto exit; } /* accept upstream's notion of segment and distribute along */ gst_segment_set_newsegment_full (&demux->segment, update, rate, arate, GST_FORMAT_TIME, start, stop, start); GST_DEBUG_OBJECT (demux, "Pushing newseg update %d, rate %g, " "applied rate %g, format %d, start %" G_GINT64_FORMAT ", " "stop %" G_GINT64_FORMAT, update, rate, arate, GST_FORMAT_TIME, start, stop); gst_qtdemux_push_event (demux, gst_event_new_new_segment_full (update, rate, arate, GST_FORMAT_TIME, start, stop, start)); /* clear leftover in current segment, if any */ gst_adapter_clear (demux->adapter); /* set up streaming thread */ gst_qtdemux_find_sample (demux, offset, TRUE, TRUE, &stream, &idx, NULL); demux->offset = offset; if (stream) { demux->todrop = stream->samples[idx].offset - offset; demux->neededbytes = demux->todrop + stream->samples[idx].size; } else { /* set up for EOS */ demux->neededbytes = -1; demux->todrop = 0; } exit: gst_event_unref (event); res = TRUE; goto drop; break; } case GST_EVENT_FLUSH_STOP: { gint i; /* clean up, force EOS if no more info follows */ gst_adapter_clear (demux->adapter); demux->offset = 0; demux->neededbytes = -1; /* reset flow return, e.g. following seek */ for (i = 0; i < demux->n_streams; i++) { demux->streams[i]->last_ret = GST_FLOW_OK; demux->streams[i]->sent_eos = FALSE; } break; } case GST_EVENT_EOS: /* If we are in push mode, and get an EOS before we've seen any streams, * then error out - we have nowhere to send the EOS */ if (!demux->pullbased && demux->n_streams == 0) { GST_ELEMENT_ERROR (demux, STREAM, DECODE, (_("This file contains no playable streams.")), ("no known streams found")); } break; default: break; } res = gst_pad_event_default (demux->sinkpad, event); drop: return res; } static GstStateChangeReturn gst_qtdemux_change_state (GstElement * element, GstStateChange transition) { GstQTDemux *qtdemux = GST_QTDEMUX (element); GstStateChangeReturn result = GST_STATE_CHANGE_FAILURE; switch (transition) { case GST_STATE_CHANGE_PAUSED_TO_READY: break; default: break; } result = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); switch (transition) { case GST_STATE_CHANGE_PAUSED_TO_READY:{ gint n; qtdemux->state = QTDEMUX_STATE_INITIAL; qtdemux->neededbytes = 16; qtdemux->todrop = 0; qtdemux->pullbased = FALSE; qtdemux->offset = 0; qtdemux->mdatoffset = GST_CLOCK_TIME_NONE; if (qtdemux->mdatbuffer) gst_buffer_unref (qtdemux->mdatbuffer); qtdemux->mdatbuffer = NULL; if (qtdemux->comp_brands) gst_buffer_unref (qtdemux->comp_brands); qtdemux->comp_brands = NULL; if (qtdemux->tag_list) gst_tag_list_free (qtdemux->tag_list); qtdemux->tag_list = NULL; gst_adapter_clear (qtdemux->adapter); for (n = 0; n < qtdemux->n_streams; n++) { QtDemuxStream *stream = qtdemux->streams[n]; while (stream->buffers) { gst_buffer_unref (GST_BUFFER_CAST (stream->buffers->data)); stream->buffers = g_slist_delete_link (stream->buffers, stream->buffers); } if (stream->pad) gst_element_remove_pad (element, stream->pad); if (stream->samples) g_free (stream->samples); if (stream->caps) gst_caps_unref (stream->caps); if (stream->segments) g_free (stream->segments); if (stream->pending_tags) gst_tag_list_free (stream->pending_tags); g_free (stream); } qtdemux->major_brand = 0; qtdemux->n_streams = 0; qtdemux->n_video_streams = 0; qtdemux->n_audio_streams = 0; qtdemux->n_sub_streams = 0; gst_segment_init (&qtdemux->segment, GST_FORMAT_TIME); break; } default: break; } return result; } static void qtdemux_parse_ftyp (GstQTDemux * qtdemux, const guint8 * buffer, gint length) { /* only consider at least a sufficiently complete ftyp atom */ if (length >= 20) { GstBuffer *buf; qtdemux->major_brand = QT_FOURCC (buffer + 8); GST_DEBUG_OBJECT (qtdemux, "major brand: %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (qtdemux->major_brand)); buf = qtdemux->comp_brands = gst_buffer_new_and_alloc (length - 16); memcpy (GST_BUFFER_DATA (buf), buffer + 16, GST_BUFFER_SIZE (buf)); } } static void extract_initial_length_and_fourcc (const guint8 * data, guint64 * plength, guint32 * pfourcc) { guint64 length; guint32 fourcc; length = QT_UINT32 (data); GST_DEBUG ("length 0x%08" G_GINT64_MODIFIER "x", length); fourcc = QT_FOURCC (data + 4); GST_DEBUG ("atom type %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (fourcc)); if (length == 0) { length = G_MAXUINT32; } else if (length == 1) { /* this means we have an extended size, which is the 64 bit value of * the next 8 bytes */ length = QT_UINT64 (data + 8); GST_DEBUG ("length 0x%08" G_GINT64_MODIFIER "x", length); } if (plength) *plength = length; if (pfourcc) *pfourcc = fourcc; } static GstFlowReturn gst_qtdemux_loop_state_header (GstQTDemux * qtdemux) { guint64 length = 0; guint32 fourcc = 0; GstBuffer *buf = NULL; GstFlowReturn ret = GST_FLOW_OK; guint64 cur_offset = qtdemux->offset; ret = gst_pad_pull_range (qtdemux->sinkpad, cur_offset, 16, &buf); if (G_UNLIKELY (ret != GST_FLOW_OK)) goto beach; if (G_LIKELY (GST_BUFFER_SIZE (buf) == 16)) extract_initial_length_and_fourcc (GST_BUFFER_DATA (buf), &length, &fourcc); gst_buffer_unref (buf); if (G_UNLIKELY (length == 0)) { GST_ELEMENT_ERROR (qtdemux, STREAM, DECODE, (_("This file is invalid and cannot be played.")), ("Header atom '%" GST_FOURCC_FORMAT "' has empty length", GST_FOURCC_ARGS (fourcc))); ret = GST_FLOW_ERROR; goto beach; } switch (fourcc) { case FOURCC_mdat: case FOURCC_free: case FOURCC_wide: case FOURCC_PICT: case FOURCC_pnot: { GST_LOG_OBJECT (qtdemux, "skipping atom '%" GST_FOURCC_FORMAT "' at %" G_GUINT64_FORMAT, GST_FOURCC_ARGS (fourcc), cur_offset); qtdemux->offset += length; break; } case FOURCC_moov: { GstBuffer *moov; ret = gst_pad_pull_range (qtdemux->sinkpad, cur_offset, length, &moov); if (ret != GST_FLOW_OK) goto beach; if (length != GST_BUFFER_SIZE (moov)) { /* Some files have a 'moov' atom at the end of the file which contains * a terminal 'free' atom where the body of the atom is missing. * Check for, and permit, this special case. */ if (GST_BUFFER_SIZE (moov) >= 8) { guint8 *final_data = GST_BUFFER_DATA (moov) + (GST_BUFFER_SIZE (moov) - 8); guint32 final_length = QT_UINT32 (final_data); guint32 final_fourcc = QT_FOURCC (final_data + 4); if (final_fourcc == FOURCC_free && GST_BUFFER_SIZE (moov) + final_length - 8 == length) { /* Ok, we've found that special case. Allocate a new buffer with * that free atom actually present. */ GstBuffer *newmoov = gst_buffer_new_and_alloc (length); gst_buffer_copy_metadata (newmoov, moov, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS | GST_BUFFER_COPY_CAPS); memcpy (GST_BUFFER_DATA (newmoov), GST_BUFFER_DATA (moov), GST_BUFFER_SIZE (moov)); memset (GST_BUFFER_DATA (newmoov) + GST_BUFFER_SIZE (moov), 0, final_length - 8); gst_buffer_unref (moov); moov = newmoov; } } } if (length != GST_BUFFER_SIZE (moov)) { GST_ELEMENT_ERROR (qtdemux, STREAM, DECODE, (_("This file is incomplete and cannot be played.")), ("We got less than expected (received %u, wanted %u, offset %" G_GUINT64_FORMAT ")", GST_BUFFER_SIZE (moov), (guint) length, cur_offset)); ret = GST_FLOW_ERROR; goto beach; } qtdemux->offset += length; qtdemux_parse_moov (qtdemux, GST_BUFFER_DATA (moov), length); qtdemux_node_dump (qtdemux, qtdemux->moov_node); qtdemux_parse_tree (qtdemux); g_node_destroy (qtdemux->moov_node); gst_buffer_unref (moov); qtdemux->moov_node = NULL; qtdemux->state = QTDEMUX_STATE_MOVIE; GST_DEBUG_OBJECT (qtdemux, "switching state to STATE_MOVIE (%d)", qtdemux->state); break; } case FOURCC_ftyp: { GstBuffer *ftyp; /* extract major brand; might come in handy for ISO vs QT issues */ ret = gst_qtdemux_pull_atom (qtdemux, cur_offset, length, &ftyp); if (ret != GST_FLOW_OK) goto beach; qtdemux->offset += length; qtdemux_parse_ftyp (qtdemux, GST_BUFFER_DATA (ftyp), GST_BUFFER_SIZE (ftyp)); gst_buffer_unref (ftyp); break; } default: { GstBuffer *unknown; GST_LOG_OBJECT (qtdemux, "unknown %08x '%" GST_FOURCC_FORMAT "' of size %" G_GUINT64_FORMAT " at %" G_GUINT64_FORMAT, fourcc, GST_FOURCC_ARGS (fourcc), length, cur_offset); ret = gst_qtdemux_pull_atom (qtdemux, cur_offset, length, &unknown); if (ret != GST_FLOW_OK) goto beach; GST_MEMDUMP ("Unknown tag", GST_BUFFER_DATA (unknown), GST_BUFFER_SIZE (unknown)); gst_buffer_unref (unknown); qtdemux->offset += length; break; } } beach: return ret; } /* Seeks to the previous keyframe of the indexed stream and * aligns other streams with respect to the keyframe timestamp * of indexed stream. Only called in case of Reverse Playback */ static GstFlowReturn gst_qtdemux_seek_to_previous_keyframe (GstQTDemux * qtdemux) { guint8 n = 0; guint32 seg_idx = 0, k_index = 0; guint64 k_pos = 0, last_stop = 0; QtDemuxSegment *seg = NULL; QtDemuxStream *ref_str = NULL; /* Now we choose an arbitrary stream, get the previous keyframe timestamp * and finally align all the other streams on that timestamp with their * respective keyframes */ for (n = 0; n < qtdemux->n_streams; n++) { QtDemuxStream *str = qtdemux->streams[n]; seg_idx = gst_qtdemux_find_segment (qtdemux, str, qtdemux->segment.last_stop); /* segment not found, continue with normal flow */ if (seg_idx == -1) continue; /* No candidate yet, take that one */ if (!ref_str) { ref_str = str; continue; } /* So that stream has a segment, we prefer video streams */ if (str->subtype == FOURCC_vide) { ref_str = str; break; } } if (G_UNLIKELY (!ref_str)) { GST_DEBUG_OBJECT (qtdemux, "couldn't find any stream"); goto eos; } if (G_UNLIKELY (!ref_str->from_sample)) { GST_DEBUG_OBJECT (qtdemux, "reached the beginning of the file"); goto eos; } /* So that stream has been playing from from_sample to to_sample. We will * get the timestamp of the previous sample and search for a keyframe before * that. For audio streams we do an arbitrary jump in the past (10 samples) */ if (ref_str->subtype == FOURCC_vide) { k_index = gst_qtdemux_find_keyframe (qtdemux, ref_str, ref_str->from_sample - 1); } else { k_index = ref_str->from_sample - 10; } /* get current segment for that stream */ seg = &ref_str->segments[ref_str->segment_index]; /* Crawl back through segments to find the one containing this I frame */ while (ref_str->samples[k_index].timestamp < seg->media_start) { GST_DEBUG_OBJECT (qtdemux, "keyframe position is out of segment %u", ref_str->segment_index); if (G_UNLIKELY (!ref_str->segment_index)) { /* Reached first segment, let's consider it's EOS */ goto eos; } ref_str->segment_index--; seg = &ref_str->segments[ref_str->segment_index]; } /* Calculate time position of the keyframe and where we should stop */ k_pos = (ref_str->samples[k_index].timestamp - seg->media_start) + seg->time; last_stop = ref_str->samples[ref_str->from_sample].timestamp; last_stop = (last_stop - seg->media_start) + seg->time; GST_DEBUG_OBJECT (qtdemux, "preferred stream played from sample %u, " "now going to sample %u (pts %" GST_TIME_FORMAT ")", ref_str->from_sample, k_index, GST_TIME_ARGS (k_pos)); /* Set last_stop with the keyframe timestamp we pushed of that stream */ gst_segment_set_last_stop (&qtdemux->segment, GST_FORMAT_TIME, last_stop); GST_DEBUG_OBJECT (qtdemux, "last_stop now is %" GST_TIME_FORMAT, GST_TIME_ARGS (last_stop)); if (G_UNLIKELY (last_stop < qtdemux->segment.start)) { GST_DEBUG_OBJECT (qtdemux, "reached the beginning of segment"); goto eos; } /* Align them all on this */ for (n = 0; n < qtdemux->n_streams; n++) { guint32 index = 0; guint64 media_start = 0, seg_time = 0; QtDemuxStream *str = qtdemux->streams[n]; seg_idx = gst_qtdemux_find_segment (qtdemux, str, k_pos); GST_DEBUG_OBJECT (qtdemux, "align segment %d", seg_idx); /* segment not found, continue with normal flow */ if (seg_idx == -1) continue; /* get segment and time in the segment */ seg = &str->segments[seg_idx]; seg_time = k_pos - seg->time; /* get the media time in the segment */ media_start = seg->media_start + seg_time; /* get the index of the sample with media time */ index = gst_qtdemux_find_index (qtdemux, str, media_start); GST_DEBUG_OBJECT (qtdemux, "sample for %" GST_TIME_FORMAT " at %u", GST_TIME_ARGS (media_start), index); /* find previous keyframe */ k_index = gst_qtdemux_find_keyframe (qtdemux, str, index); /* Remember until where we want to go */ str->to_sample = str->from_sample - 1; /* Define our time position */ str->time_position = (str->samples[k_index].timestamp - seg->media_start) + seg->time; /* Now seek back in time */ gst_qtdemux_move_stream (qtdemux, str, k_index); GST_DEBUG_OBJECT (qtdemux, "keyframe at %u, time position %" GST_TIME_FORMAT " playing from sample %u to %u", k_index, GST_TIME_ARGS (str->time_position), str->from_sample, str->to_sample); } return GST_FLOW_OK; eos: return GST_FLOW_UNEXPECTED; } /* activate the given segment number @seg_idx of @stream at time @offset. * @offset is an absolute global position over all the segments. * * This will push out a NEWSEGMENT event with the right values and * position the stream index to the first decodable sample before * @offset. */ static gboolean gst_qtdemux_activate_segment (GstQTDemux * qtdemux, QtDemuxStream * stream, guint32 seg_idx, guint64 offset) { GstEvent *event; QtDemuxSegment *segment; guint32 index, kf_index; guint64 seg_time; guint64 start, stop, time; gdouble rate; GST_LOG_OBJECT (qtdemux, "activate segment %d, offset %" G_GUINT64_FORMAT, seg_idx, offset); /* update the current segment */ stream->segment_index = seg_idx; /* get the segment */ segment = &stream->segments[seg_idx]; if (G_UNLIKELY (offset < segment->time)) { GST_WARNING_OBJECT (qtdemux, "offset < segment->time %" G_GUINT64_FORMAT, segment->time); return FALSE; } /* get time in this segment */ seg_time = offset - segment->time; GST_LOG_OBJECT (qtdemux, "seg_time %" GST_TIME_FORMAT, GST_TIME_ARGS (seg_time)); if (G_UNLIKELY (seg_time > segment->duration)) { GST_LOG_OBJECT (qtdemux, "seg_time > segment->duration %" GST_TIME_FORMAT, GST_TIME_ARGS (segment->duration)); return FALSE; } /* qtdemux->segment.stop is in outside-time-realm, whereas * segment->media_stop is in track-time-realm. * * In order to compare the two, we need to bring segment.stop * into the track-time-realm */ if (qtdemux->segment.stop == -1) stop = segment->media_stop; else stop = MIN (segment->media_stop, qtdemux->segment.stop - segment->time + segment->media_start); if (qtdemux->segment.rate >= 0) { start = MIN (segment->media_start + seg_time, stop); time = offset; } else { start = segment->media_start; stop = MIN (segment->media_start + seg_time, stop); time = segment->time; } GST_DEBUG_OBJECT (qtdemux, "newsegment %d from %" GST_TIME_FORMAT " to %" GST_TIME_FORMAT ", time %" GST_TIME_FORMAT, seg_idx, GST_TIME_ARGS (start), GST_TIME_ARGS (stop), GST_TIME_ARGS (time)); /* combine global rate with that of the segment */ rate = segment->rate * qtdemux->segment.rate; /* update the segment values used for clipping */ gst_segment_init (&stream->segment, GST_FORMAT_TIME); gst_segment_set_newsegment (&stream->segment, FALSE, rate, GST_FORMAT_TIME, start, stop, time); /* now prepare and send the segment */ if (stream->pad) { event = gst_event_new_new_segment (FALSE, rate, GST_FORMAT_TIME, start, stop, time); gst_pad_push_event (stream->pad, event); /* assume we can send more data now */ stream->last_ret = GST_FLOW_OK; /* clear to send tags on this pad now */ gst_qtdemux_push_tags (qtdemux, stream); } /* and move to the keyframe before the indicated media time of the * segment */ if (qtdemux->segment.rate >= 0) { index = gst_qtdemux_find_index (qtdemux, stream, start); stream->to_sample = stream->n_samples; GST_DEBUG_OBJECT (qtdemux, "moving data pointer to %" GST_TIME_FORMAT ", index: %u, pts %" GST_TIME_FORMAT, GST_TIME_ARGS (start), index, GST_TIME_ARGS (stream->samples[index].timestamp)); } else { index = gst_qtdemux_find_index (qtdemux, stream, stop); stream->to_sample = index; GST_DEBUG_OBJECT (qtdemux, "moving data pointer to %" GST_TIME_FORMAT ", index: %u, pts %" GST_TIME_FORMAT, GST_TIME_ARGS (stop), index, GST_TIME_ARGS (stream->samples[index].timestamp)); } /* we're at the right spot */ if (index == stream->sample_index) { GST_DEBUG_OBJECT (qtdemux, "we are at the right index"); return TRUE; } /* find keyframe of the target index */ kf_index = gst_qtdemux_find_keyframe (qtdemux, stream, index); /* if we move forwards, we don't have to go back to the previous * keyframe since we already sent that. We can also just jump to * the keyframe right before the target index if there is one. */ if (index > stream->sample_index) { /* moving forwards check if we move past a keyframe */ if (kf_index > stream->sample_index) { GST_DEBUG_OBJECT (qtdemux, "moving forwards to keyframe at %u (pts %" GST_TIME_FORMAT, kf_index, GST_TIME_ARGS (stream->samples[kf_index].timestamp)); gst_qtdemux_move_stream (qtdemux, stream, kf_index); } else { GST_DEBUG_OBJECT (qtdemux, "moving forwards, keyframe at %u (pts %" GST_TIME_FORMAT " already sent", kf_index, GST_TIME_ARGS (stream->samples[kf_index].timestamp)); } } else { GST_DEBUG_OBJECT (qtdemux, "moving backwards to keyframe at %u (pts %" GST_TIME_FORMAT, kf_index, GST_TIME_ARGS (stream->samples[kf_index].timestamp)); gst_qtdemux_move_stream (qtdemux, stream, kf_index); } return TRUE; } /* prepare to get the current sample of @stream, getting essential values. * * This function will also prepare and send the segment when needed. * * Return FALSE if the stream is EOS. */ static gboolean gst_qtdemux_prepare_current_sample (GstQTDemux * qtdemux, QtDemuxStream * stream, guint64 * offset, guint * size, guint64 * timestamp, guint64 * duration, gboolean * keyframe) { QtDemuxSample *sample; guint64 time_position; guint32 seg_idx; g_return_val_if_fail (stream != NULL, FALSE); time_position = stream->time_position; if (G_UNLIKELY (time_position == -1)) goto eos; seg_idx = stream->segment_index; if (G_UNLIKELY (seg_idx == -1)) { /* find segment corresponding to time_position if we are looking * for a segment. */ seg_idx = gst_qtdemux_find_segment (qtdemux, stream, time_position); /* nothing found, we're really eos */ if (seg_idx == -1) goto eos; } /* different segment, activate it, sample_index will be set. */ if (G_UNLIKELY (stream->segment_index != seg_idx)) gst_qtdemux_activate_segment (qtdemux, stream, seg_idx, time_position); GST_LOG_OBJECT (qtdemux, "segment active, index = %u of %u", stream->sample_index, stream->n_samples); if (G_UNLIKELY (stream->sample_index >= stream->n_samples)) goto eos; /* now get the info for the sample we're at */ sample = &stream->samples[stream->sample_index]; *timestamp = sample->timestamp + sample->pts_offset; *offset = sample->offset; *size = sample->size; *duration = sample->duration; *keyframe = stream->all_keyframe || sample->keyframe; return TRUE; /* special cases */ eos: { stream->time_position = -1; return FALSE; } } /* move to the next sample in @stream. * * Moves to the next segment when needed. */ static void gst_qtdemux_advance_sample (GstQTDemux * qtdemux, QtDemuxStream * stream) { QtDemuxSample *sample; QtDemuxSegment *segment; if (G_UNLIKELY (stream->sample_index >= stream->to_sample)) { /* Mark the stream as EOS */ GST_DEBUG_OBJECT (qtdemux, "reached max allowed sample %u, mark EOS", stream->to_sample); stream->time_position = -1; return; } /* move to next sample */ stream->sample_index++; /* get current segment */ segment = &stream->segments[stream->segment_index]; /* reached the last sample, we need the next segment */ if (G_UNLIKELY (stream->sample_index >= stream->n_samples)) goto next_segment; /* get next sample */ sample = &stream->samples[stream->sample_index]; /* see if we are past the segment */ if (G_UNLIKELY (sample->timestamp >= segment->media_stop)) goto next_segment; if (sample->timestamp >= segment->media_start) { /* inside the segment, update time_position, looks very familiar to * GStreamer segments, doesn't it? */ stream->time_position = (sample->timestamp - segment->media_start) + segment->time; } else { /* not yet in segment, time does not yet increment. This means * that we are still prerolling keyframes to the decoder so it can * decode the first sample of the segment. */ stream->time_position = segment->time; } return; /* move to the next segment */ next_segment: { GST_DEBUG_OBJECT (qtdemux, "segment %d ended ", stream->segment_index); if (stream->segment_index == stream->n_segments - 1) { /* are we at the end of the last segment, we're EOS */ stream->time_position = -1; } else { /* else we're only at the end of the current segment */ stream->time_position = segment->stop_time; } /* make sure we select a new segment */ stream->segment_index = -1; } } static void gst_qtdemux_sync_streams (GstQTDemux * demux) { gint i; if (demux->n_streams <= 1) return; for (i = 0; i < demux->n_streams; i++) { QtDemuxStream *stream; GstClockTime end_time; stream = demux->streams[i]; if (!stream->pad) continue; /* TODO advance time on subtitle streams here, if any some day */ /* some clips/trailers may have unbalanced streams at the end, * so send EOS on shorter stream to prevent stalling others */ /* do not mess with EOS if SEGMENT seeking */ if (demux->segment.flags & GST_SEEK_FLAG_SEGMENT) continue; if (demux->pullbased) { /* loop mode is sample time based */ if (stream->time_position != -1) continue; } else { /* push mode is byte position based */ if (stream->samples[stream->n_samples - 1].offset >= demux->offset) continue; } if (stream->sent_eos) continue; /* only act if some gap */ end_time = stream->segments[stream->n_segments - 1].stop_time; GST_LOG_OBJECT (demux, "current position: %" GST_TIME_FORMAT ", stream end: %" GST_TIME_FORMAT, GST_TIME_ARGS (end_time), GST_TIME_ARGS (demux->segment.last_stop)); if (end_time + 2 * GST_SECOND < demux->segment.last_stop) { GST_DEBUG_OBJECT (demux, "sending EOS for stream %s", GST_PAD_NAME (stream->pad)); stream->sent_eos = TRUE; gst_pad_push_event (stream->pad, gst_event_new_eos ()); } } } /* UNEXPECTED and NOT_LINKED need to be combined. This means that we return: * * GST_FLOW_NOT_LINKED: when all pads NOT_LINKED. * GST_FLOW_UNEXPECTED: when all pads UNEXPECTED or NOT_LINKED. */ static GstFlowReturn gst_qtdemux_combine_flows (GstQTDemux * demux, QtDemuxStream * stream, GstFlowReturn ret) { gint i; gboolean unexpected = FALSE, not_linked = TRUE; GST_LOG_OBJECT (demux, "flow return: %s", gst_flow_get_name (ret)); /* store the value */ stream->last_ret = ret; for (i = 0; i < demux->n_streams; i++) { QtDemuxStream *ostream = demux->streams[i]; ret = ostream->last_ret; /* no unexpected or unlinked, return */ if (G_LIKELY (ret != GST_FLOW_UNEXPECTED && ret != GST_FLOW_NOT_LINKED)) goto done; /* we check to see if we have at least 1 unexpected or all unlinked */ unexpected |= (ret == GST_FLOW_UNEXPECTED); not_linked &= (ret == GST_FLOW_NOT_LINKED); } /* when we get here, we all have unlinked or unexpected */ if (not_linked) ret = GST_FLOW_NOT_LINKED; else if (unexpected) ret = GST_FLOW_UNEXPECTED; done: GST_LOG_OBJECT (demux, "combined flow return: %s", gst_flow_get_name (ret)); return ret; } /* the input buffer metadata must be writable. Returns NULL when the buffer is * completely cliped */ static GstBuffer * gst_qtdemux_clip_buffer (GstQTDemux * qtdemux, QtDemuxStream * stream, GstBuffer * buf) { gint64 start, stop, cstart, cstop, diff; GstClockTime timestamp = GST_CLOCK_TIME_NONE, duration = GST_CLOCK_TIME_NONE; guint8 *data; guint size; gint num_rate, denom_rate; gint frame_size; gboolean clip_data; data = GST_BUFFER_DATA (buf); size = GST_BUFFER_SIZE (buf); /* depending on the type, setup the clip parameters */ if (stream->subtype == FOURCC_soun) { frame_size = stream->bytes_per_frame; num_rate = GST_SECOND; denom_rate = (gint) stream->rate; clip_data = TRUE; } else if (stream->subtype == FOURCC_vide) { frame_size = size; num_rate = stream->fps_n; denom_rate = stream->fps_d; clip_data = FALSE; } else goto wrong_type; /* we can only clip if we have a valid timestamp */ timestamp = GST_BUFFER_TIMESTAMP (buf); if (G_UNLIKELY (!GST_CLOCK_TIME_IS_VALID (timestamp))) goto no_timestamp; if (G_LIKELY (GST_BUFFER_DURATION_IS_VALID (buf))) { duration = GST_BUFFER_DURATION (buf); } else { duration = gst_util_uint64_scale_int (size / frame_size, num_rate, denom_rate); } start = timestamp; stop = start + duration; if (G_UNLIKELY (!gst_segment_clip (&stream->segment, GST_FORMAT_TIME, start, stop, &cstart, &cstop))) goto clipped; /* see if some clipping happened */ diff = cstart - start; if (diff > 0) { timestamp = cstart; duration -= diff; if (clip_data) { /* bring clipped time to samples and to bytes */ diff = gst_util_uint64_scale_int (diff, denom_rate, num_rate); diff *= frame_size; GST_DEBUG_OBJECT (qtdemux, "clipping start to %" GST_TIME_FORMAT " %" G_GUINT64_FORMAT " bytes", GST_TIME_ARGS (cstart), diff); data += diff; size -= diff; } } diff = stop - cstop; if (diff > 0) { duration -= diff; if (clip_data) { /* bring clipped time to samples and then to bytes */ diff = gst_util_uint64_scale_int (diff, denom_rate, num_rate); diff *= frame_size; GST_DEBUG_OBJECT (qtdemux, "clipping stop to %" GST_TIME_FORMAT " %" G_GUINT64_FORMAT " bytes", GST_TIME_ARGS (cstop), diff); size -= diff; } } GST_BUFFER_TIMESTAMP (buf) = timestamp; GST_BUFFER_DURATION (buf) = duration; GST_BUFFER_SIZE (buf) = size; GST_BUFFER_DATA (buf) = data; return buf; /* dropped buffer */ wrong_type: { GST_DEBUG_OBJECT (qtdemux, "unknown stream type"); return buf; } no_timestamp: { GST_DEBUG_OBJECT (qtdemux, "no timestamp on buffer"); return buf; } clipped: { GST_DEBUG_OBJECT (qtdemux, "clipped buffer"); gst_buffer_unref (buf); return NULL; } } /* the input buffer metadata must be writable, * but time/duration etc not yet set and need not be preserved */ static GstBuffer * gst_qtdemux_process_buffer (GstQTDemux * qtdemux, QtDemuxStream * stream, GstBuffer * buf) { guint8 *data; guint size, nsize = 0; gchar *str; data = GST_BUFFER_DATA (buf); size = GST_BUFFER_SIZE (buf); /* not many cases for now */ if (G_UNLIKELY (stream->fourcc == FOURCC_mp4s)) { /* send a one time dvd clut event */ if (stream->pending_event && stream->pad) gst_pad_push_event (stream->pad, stream->pending_event); stream->pending_event = NULL; /* no further processing needed */ stream->need_process = FALSE; } if (G_UNLIKELY (stream->fourcc != FOURCC_tx3g)) { return buf; } if (G_LIKELY (size >= 2)) { nsize = GST_READ_UINT16_BE (data); nsize = MIN (nsize, size - 2); } GST_LOG_OBJECT (qtdemux, "3GPP timed text subtitle: %d/%d", nsize, size); /* takes care of UTF-8 validation or UTF-16 recognition, * no other encoding expected */ str = gst_tag_freeform_string_to_utf8 ((gchar *) data + 2, nsize, NULL); if (str) { gst_buffer_unref (buf); buf = gst_buffer_new (); GST_BUFFER_DATA (buf) = GST_BUFFER_MALLOCDATA (buf) = (guint8 *) str; GST_BUFFER_SIZE (buf) = strlen (str); } else { /* may be 0-size subtitle, which is also sent to keep pipeline going */ GST_BUFFER_DATA (buf) = data + 2; GST_BUFFER_SIZE (buf) = nsize; } /* FIXME ? convert optional subsequent style info to markup */ return buf; } /* Sets a buffer's attributes properly and pushes it downstream. * Also checks for additional actions and custom processing that may * need to be done first. */ static gboolean gst_qtdemux_decorate_and_push_buffer (GstQTDemux * qtdemux, QtDemuxStream * stream, GstBuffer * buf, guint64 timestamp, guint64 duration, gboolean keyframe, guint64 position) { GstFlowReturn ret = GST_FLOW_OK; if (G_UNLIKELY (stream->fourcc == FOURCC_rtsp)) { GstMessage *m; gchar *url; url = g_strndup ((gchar *) GST_BUFFER_DATA (buf), GST_BUFFER_SIZE (buf)); /* we have RTSP redirect now */ m = gst_message_new_element (GST_OBJECT_CAST (qtdemux), gst_structure_new ("redirect", "new-location", G_TYPE_STRING, url, NULL)); g_free (url); gst_element_post_message (GST_ELEMENT_CAST (qtdemux), m); } /* position reporting */ if (qtdemux->segment.rate >= 0) { gst_segment_set_last_stop (&qtdemux->segment, GST_FORMAT_TIME, position); gst_qtdemux_sync_streams (qtdemux); } if (G_UNLIKELY (!stream->pad)) { GST_DEBUG_OBJECT (qtdemux, "No output pad for stream, ignoring"); gst_buffer_unref (buf); goto exit; } /* send out pending buffers */ while (stream->buffers) { GstBuffer *buffer = (GstBuffer *) stream->buffers->data; if (G_UNLIKELY (stream->discont)) { GST_LOG_OBJECT (qtdemux, "marking discont buffer"); GST_BUFFER_FLAG_SET (buffer, GST_BUFFER_FLAG_DISCONT); stream->discont = FALSE; } gst_buffer_set_caps (buffer, stream->caps); gst_pad_push (stream->pad, buffer); stream->buffers = g_slist_delete_link (stream->buffers, stream->buffers); } /* we're going to modify the metadata */ buf = gst_buffer_make_metadata_writable (buf); if (G_UNLIKELY (stream->need_process)) buf = gst_qtdemux_process_buffer (qtdemux, stream, buf); GST_BUFFER_TIMESTAMP (buf) = timestamp; GST_BUFFER_DURATION (buf) = duration; GST_BUFFER_OFFSET (buf) = -1; GST_BUFFER_OFFSET_END (buf) = -1; if (G_UNLIKELY (stream->padding)) { GST_BUFFER_DATA (buf) += stream->padding; GST_BUFFER_SIZE (buf) -= stream->padding; } if (stream->need_clip) buf = gst_qtdemux_clip_buffer (qtdemux, stream, buf); if (G_UNLIKELY (buf == NULL)) goto exit; if (G_UNLIKELY (stream->discont)) { GST_LOG_OBJECT (qtdemux, "marking discont buffer"); GST_BUFFER_FLAG_SET (buf, GST_BUFFER_FLAG_DISCONT); stream->discont = FALSE; } if (!keyframe) GST_BUFFER_FLAG_SET (buf, GST_BUFFER_FLAG_DELTA_UNIT); gst_buffer_set_caps (buf, stream->caps); GST_LOG_OBJECT (qtdemux, "Pushing buffer with time %" GST_TIME_FORMAT ", duration %" GST_TIME_FORMAT " on pad %s", GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (buf)), GST_TIME_ARGS (GST_BUFFER_DURATION (buf)), GST_PAD_NAME (stream->pad)); ret = gst_pad_push (stream->pad, buf); exit: return ret; } static GstFlowReturn gst_qtdemux_loop_state_movie (GstQTDemux * qtdemux) { GstFlowReturn ret = GST_FLOW_OK; GstBuffer *buf = NULL; QtDemuxStream *stream; guint64 min_time; guint64 offset = 0; guint64 timestamp = GST_CLOCK_TIME_NONE; guint64 duration = 0; gboolean keyframe = FALSE; guint size = 0; gint index; gint i; gst_qtdemux_push_pending_newsegment (qtdemux); /* Figure out the next stream sample to output, min_time is expressed in * global time and runs over the edit list segments. */ min_time = G_MAXUINT64; index = -1; for (i = 0; i < qtdemux->n_streams; i++) { guint64 position; stream = qtdemux->streams[i]; position = stream->time_position; /* position of -1 is EOS */ if (position != -1 && position < min_time) { min_time = position; index = i; } } /* all are EOS */ if (G_UNLIKELY (index == -1)) { GST_DEBUG_OBJECT (qtdemux, "all streams are EOS"); goto eos; } /* check for segment end */ if (G_UNLIKELY (qtdemux->segment.stop != -1 && qtdemux->segment.stop < min_time)) { GST_DEBUG_OBJECT (qtdemux, "we reached the end of our segment."); goto eos; } stream = qtdemux->streams[index]; /* fetch info for the current sample of this stream */ if (G_UNLIKELY (!gst_qtdemux_prepare_current_sample (qtdemux, stream, &offset, &size, ×tamp, &duration, &keyframe))) goto eos_stream; GST_LOG_OBJECT (qtdemux, "pushing from stream %d, offset %" G_GUINT64_FORMAT ", size %d, timestamp=%" GST_TIME_FORMAT ", duration %" GST_TIME_FORMAT, index, offset, size, GST_TIME_ARGS (timestamp), GST_TIME_ARGS (duration)); /* hmm, empty sample, skip and move to next sample */ if (G_UNLIKELY (size <= 0)) goto next; /* last pushed sample was out of boundary, goto next sample */ if (G_UNLIKELY (stream->last_ret == GST_FLOW_UNEXPECTED)) goto next; GST_LOG_OBJECT (qtdemux, "reading %d bytes @ %" G_GUINT64_FORMAT, size, offset); ret = gst_qtdemux_pull_atom (qtdemux, offset, size, &buf); if (G_UNLIKELY (ret != GST_FLOW_OK)) goto beach; ret = gst_qtdemux_decorate_and_push_buffer (qtdemux, stream, buf, timestamp, duration, keyframe, min_time); /* combine flows */ ret = gst_qtdemux_combine_flows (qtdemux, stream, ret); /* ignore unlinked, we will not push on the pad anymore and we will EOS when * we have no more data for the pad to push */ if (ret == GST_FLOW_UNEXPECTED) ret = GST_FLOW_OK; next: gst_qtdemux_advance_sample (qtdemux, stream); beach: return ret; /* special cases */ eos: { GST_DEBUG_OBJECT (qtdemux, "No samples left for any streams - EOS"); ret = GST_FLOW_UNEXPECTED; goto beach; } eos_stream: { GST_DEBUG_OBJECT (qtdemux, "No samples left for stream"); /* EOS will be raised if all are EOS */ ret = GST_FLOW_OK; goto beach; } } static void gst_qtdemux_loop (GstPad * pad) { GstQTDemux *qtdemux; guint64 cur_offset; GstFlowReturn ret; qtdemux = GST_QTDEMUX (gst_pad_get_parent (pad)); cur_offset = qtdemux->offset; GST_LOG_OBJECT (qtdemux, "loop at position %" G_GUINT64_FORMAT ", state %d", cur_offset, qtdemux->state); switch (qtdemux->state) { case QTDEMUX_STATE_INITIAL: case QTDEMUX_STATE_HEADER: ret = gst_qtdemux_loop_state_header (qtdemux); break; case QTDEMUX_STATE_MOVIE: ret = gst_qtdemux_loop_state_movie (qtdemux); if (qtdemux->segment.rate < 0 && ret == GST_FLOW_UNEXPECTED) { ret = gst_qtdemux_seek_to_previous_keyframe (qtdemux); } break; default: /* ouch */ goto invalid_state; } /* if something went wrong, pause */ if (ret != GST_FLOW_OK) goto pause; done: gst_object_unref (qtdemux); return; /* ERRORS */ invalid_state: { GST_ELEMENT_ERROR (qtdemux, STREAM, FAILED, (NULL), ("streaming stopped, invalid state")); qtdemux->segment_running = FALSE; gst_pad_pause_task (pad); gst_qtdemux_push_event (qtdemux, gst_event_new_eos ()); goto done; } pause: { const gchar *reason = gst_flow_get_name (ret); GST_LOG_OBJECT (qtdemux, "pausing task, reason %s", reason); qtdemux->segment_running = FALSE; gst_pad_pause_task (pad); /* fatal errors need special actions */ if (GST_FLOW_IS_FATAL (ret) || ret == GST_FLOW_NOT_LINKED) { /* check EOS */ if (ret == GST_FLOW_UNEXPECTED) { if (qtdemux->n_streams == 0) { /* we have no streams, post an error */ GST_ELEMENT_ERROR (qtdemux, STREAM, DECODE, (_("This file contains no playable streams.")), ("no known streams found")); } if (qtdemux->segment.flags & GST_SEEK_FLAG_SEGMENT) { gint64 stop; /* FIXME: I am not sure this is the right fix. If the sinks are * supposed to detect the segment is complete and accumulate * automatically, it does not seem to work here. Need more work */ qtdemux->segment_running = TRUE; if ((stop = qtdemux->segment.stop) == -1) stop = qtdemux->segment.duration; if (qtdemux->segment.rate >= 0) { GST_LOG_OBJECT (qtdemux, "Sending segment done, at end of segment"); gst_element_post_message (GST_ELEMENT_CAST (qtdemux), gst_message_new_segment_done (GST_OBJECT_CAST (qtdemux), GST_FORMAT_TIME, stop)); } else { /* For Reverse Playback */ GST_LOG_OBJECT (qtdemux, "Sending segment done, at start of segment"); gst_element_post_message (GST_ELEMENT_CAST (qtdemux), gst_message_new_segment_done (GST_OBJECT_CAST (qtdemux), GST_FORMAT_TIME, qtdemux->segment.start)); } } else { GST_LOG_OBJECT (qtdemux, "Sending EOS at end of segment"); gst_qtdemux_push_event (qtdemux, gst_event_new_eos ()); } } else { GST_ELEMENT_ERROR (qtdemux, STREAM, FAILED, (NULL), ("streaming stopped, reason %s", reason)); gst_qtdemux_push_event (qtdemux, gst_event_new_eos ()); } } goto done; } } /* * next_entry_size * * Returns the size of the first entry at the current offset. * If -1, there are none (which means EOS or empty file). */ static guint64 next_entry_size (GstQTDemux * demux) { QtDemuxStream *stream; int i; int smallidx = -1; guint64 smalloffs = (guint64) - 1; QtDemuxSample *sample; GST_LOG_OBJECT (demux, "Finding entry at offset %" G_GUINT64_FORMAT, demux->offset); for (i = 0; i < demux->n_streams; i++) { stream = demux->streams[i]; if (stream->sample_index == -1) stream->sample_index = 0; if (stream->sample_index >= stream->n_samples) { GST_LOG_OBJECT (demux, "stream %d samples exhausted", i); continue; } sample = &stream->samples[stream->sample_index]; GST_LOG_OBJECT (demux, "Checking Stream %d (sample_index:%d / offset:%" G_GUINT64_FORMAT " / size:%" G_GUINT32_FORMAT ")", i, stream->sample_index, sample->offset, sample->size); if (((smalloffs == -1) || (sample->offset < smalloffs)) && (sample->size)) { smallidx = i; smalloffs = sample->offset; } } GST_LOG_OBJECT (demux, "stream %d offset %" G_GUINT64_FORMAT " demux->offset :%" G_GUINT64_FORMAT, smallidx, smalloffs, demux->offset); if (smallidx == -1) return -1; stream = demux->streams[smallidx]; sample = &stream->samples[stream->sample_index]; if (sample->offset >= demux->offset) { demux->todrop = sample->offset - demux->offset; return sample->size + demux->todrop; } GST_DEBUG_OBJECT (demux, "There wasn't any entry at offset %" G_GUINT64_FORMAT, demux->offset); return -1; } static void gst_qtdemux_post_progress (GstQTDemux * demux, gint num, gint denom) { gint perc = (gint) ((gdouble) num * 100.0 / (gdouble) denom); gst_element_post_message (GST_ELEMENT_CAST (demux), gst_message_new_element (GST_OBJECT_CAST (demux), gst_structure_new ("progress", "percent", G_TYPE_INT, perc, NULL))); } /* FIXME, unverified after edit list updates */ static GstFlowReturn gst_qtdemux_chain (GstPad * sinkpad, GstBuffer * inbuf) { GstQTDemux *demux; GstFlowReturn ret = GST_FLOW_OK; demux = GST_QTDEMUX (gst_pad_get_parent (sinkpad)); gst_adapter_push (demux->adapter, inbuf); /* we never really mean to buffer that much */ if (demux->neededbytes == -1) goto eos; GST_DEBUG_OBJECT (demux, "pushing in inbuf %p, neededbytes:%u, available:%u", inbuf, demux->neededbytes, gst_adapter_available (demux->adapter)); while (((gst_adapter_available (demux->adapter)) >= demux->neededbytes) && (ret == GST_FLOW_OK)) { GST_DEBUG_OBJECT (demux, "state:%d , demux->neededbytes:%d, demux->offset:%" G_GUINT64_FORMAT, demux->state, demux->neededbytes, demux->offset); switch (demux->state) { case QTDEMUX_STATE_INITIAL:{ const guint8 *data; guint32 fourcc; guint64 size; /* prepare newsegment to send when streaming actually starts */ if (!demux->pending_newsegment) { demux->pending_newsegment = gst_event_new_new_segment (FALSE, 1.0, GST_FORMAT_TIME, 0, GST_CLOCK_TIME_NONE, 0); } data = gst_adapter_peek (demux->adapter, demux->neededbytes); /* get fourcc/length, set neededbytes */ extract_initial_length_and_fourcc ((guint8 *) data, &size, &fourcc); GST_DEBUG_OBJECT (demux, "Peeking found [%" GST_FOURCC_FORMAT "] " "size: %" G_GUINT64_FORMAT, GST_FOURCC_ARGS (fourcc), size); if (size == 0) { GST_ELEMENT_ERROR (demux, STREAM, DECODE, (_("This file is invalid and cannot be played.")), ("initial atom '%" GST_FOURCC_FORMAT "' has empty length", GST_FOURCC_ARGS (fourcc))); ret = GST_FLOW_ERROR; break; } if (fourcc == FOURCC_mdat) { if (demux->n_streams > 0) { demux->state = QTDEMUX_STATE_MOVIE; demux->neededbytes = next_entry_size (demux); } else { guint bs; buffer_data: /* there may be multiple mdat (or alike) buffers */ /* sanity check */ if (demux->mdatbuffer) bs = GST_BUFFER_SIZE (demux->mdatbuffer); else bs = 0; if (size + bs > 10 * (1 << 20)) goto no_moov; demux->state = QTDEMUX_STATE_BUFFER_MDAT; demux->neededbytes = size; if (!demux->mdatbuffer) demux->mdatoffset = demux->offset; } } else if (G_UNLIKELY (size > QTDEMUX_MAX_ATOM_SIZE)) { GST_ELEMENT_ERROR (demux, STREAM, DECODE, (_("This file is invalid and cannot be played.")), ("atom %" GST_FOURCC_FORMAT " has bogus size %" G_GUINT64_FORMAT, GST_FOURCC_ARGS (fourcc), size)); ret = GST_FLOW_ERROR; break; } else { /* this means we already started buffering and still no moov header, * let's continue buffering everything till we get moov */ if (demux->mdatbuffer && (fourcc != FOURCC_moov)) goto buffer_data; demux->neededbytes = size; demux->state = QTDEMUX_STATE_HEADER; } break; } case QTDEMUX_STATE_HEADER:{ const guint8 *data; guint32 fourcc; GST_DEBUG_OBJECT (demux, "In header"); data = gst_adapter_peek (demux->adapter, demux->neededbytes); /* parse the header */ extract_initial_length_and_fourcc (data, NULL, &fourcc); if (fourcc == FOURCC_moov) { GST_DEBUG_OBJECT (demux, "Parsing [moov]"); qtdemux_parse_moov (demux, data, demux->neededbytes); qtdemux_node_dump (demux, demux->moov_node); qtdemux_parse_tree (demux); g_node_destroy (demux->moov_node); demux->moov_node = NULL; GST_DEBUG_OBJECT (demux, "Finished parsing the header"); } else if (fourcc == FOURCC_ftyp) { GST_DEBUG_OBJECT (demux, "Parsing [ftyp]"); qtdemux_parse_ftyp (demux, data, demux->neededbytes); } else { GST_WARNING_OBJECT (demux, "Unknown fourcc while parsing header : %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (fourcc)); /* Let's jump that one and go back to initial state */ } if (demux->mdatbuffer && demux->n_streams) { GstBuffer *buf; /* the mdat was before the header */ GST_DEBUG_OBJECT (demux, "We have n_streams:%d and mdatbuffer:%p", demux->n_streams, demux->mdatbuffer); /* restore our adapter/offset view of things with upstream; * put preceding buffered data ahead of current moov data. * This should also handle evil mdat, moov, mdat cases and alike */ buf = gst_adapter_take_buffer (demux->adapter, gst_adapter_available (demux->adapter)); gst_adapter_clear (demux->adapter); gst_adapter_push (demux->adapter, demux->mdatbuffer); gst_adapter_push (demux->adapter, buf); demux->mdatbuffer = NULL; demux->offset = demux->mdatoffset; demux->neededbytes = next_entry_size (demux); demux->state = QTDEMUX_STATE_MOVIE; } else { GST_DEBUG_OBJECT (demux, "Carrying on normally"); gst_adapter_flush (demux->adapter, demux->neededbytes); demux->offset += demux->neededbytes; demux->neededbytes = 16; demux->state = QTDEMUX_STATE_INITIAL; } break; } case QTDEMUX_STATE_BUFFER_MDAT:{ GstBuffer *buf; GST_DEBUG_OBJECT (demux, "Got our buffer at offset %" G_GUINT64_FORMAT, demux->offset); buf = gst_adapter_take_buffer (demux->adapter, demux->neededbytes); GST_DEBUG_OBJECT (demux, "mdatbuffer starts with %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (QT_FOURCC (GST_BUFFER_DATA (buf) + 4))); if (demux->mdatbuffer) demux->mdatbuffer = gst_buffer_join (demux->mdatbuffer, buf); else demux->mdatbuffer = buf; demux->offset += demux->neededbytes; demux->neededbytes = 16; demux->state = QTDEMUX_STATE_INITIAL; gst_qtdemux_post_progress (demux, 1, 1); break; } case QTDEMUX_STATE_MOVIE:{ GstBuffer *outbuf; QtDemuxStream *stream = NULL; int i = -1; guint64 timestamp, duration, position; gboolean keyframe; GST_DEBUG_OBJECT (demux, "BEGIN // in MOVIE for offset %" G_GUINT64_FORMAT, demux->offset); if (demux->todrop) { GST_LOG_OBJECT (demux, "Dropping %d bytes", demux->todrop); gst_adapter_flush (demux->adapter, demux->todrop); demux->neededbytes -= demux->todrop; demux->offset += demux->todrop; } /* first buffer? */ /* initial newsegment sent here after having added pads, * possible others in sink_event */ if (G_UNLIKELY (demux->pending_newsegment)) { gst_qtdemux_push_event (demux, demux->pending_newsegment); demux->pending_newsegment = NULL; /* clear to send tags on all streams */ for (i = 0; i < demux->n_streams; i++) { gst_qtdemux_push_tags (demux, demux->streams[i]); } } /* Figure out which stream this is packet belongs to */ for (i = 0; i < demux->n_streams; i++) { stream = demux->streams[i]; if (stream->sample_index >= stream->n_samples) continue; GST_LOG_OBJECT (demux, "Checking stream %d (sample_index:%d / offset:%" G_GUINT64_FORMAT " / size:%d)", i, stream->sample_index, stream->samples[stream->sample_index].offset, stream->samples[stream->sample_index].size); if (stream->samples[stream->sample_index].offset == demux->offset) break; } if (G_UNLIKELY (stream == NULL || i == demux->n_streams)) goto unknown_stream; /* Put data in a buffer, set timestamps, caps, ... */ outbuf = gst_adapter_take_buffer (demux->adapter, demux->neededbytes); GST_DEBUG_OBJECT (demux, "stream : %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (stream->fourcc)); g_return_val_if_fail (outbuf != NULL, GST_FLOW_ERROR); position = stream->samples[stream->sample_index].timestamp; timestamp = position + stream->samples[stream->sample_index].pts_offset; duration = stream->samples[stream->sample_index].duration; keyframe = stream->all_keyframe || stream->samples[stream->sample_index].keyframe; ret = gst_qtdemux_decorate_and_push_buffer (demux, stream, outbuf, timestamp, duration, keyframe, position); /* combine flows */ ret = gst_qtdemux_combine_flows (demux, stream, ret); stream->sample_index++; /* update current offset and figure out size of next buffer */ GST_LOG_OBJECT (demux, "increasing offset %" G_GUINT64_FORMAT " by %u", demux->offset, demux->neededbytes); demux->offset += demux->neededbytes; GST_LOG_OBJECT (demux, "offset is now %" G_GUINT64_FORMAT, demux->offset); if ((demux->neededbytes = next_entry_size (demux)) == -1) goto eos; break; } default: goto invalid_state; } } /* when buffering movie data, at least show user something is happening */ if (ret == GST_FLOW_OK && demux->state == QTDEMUX_STATE_BUFFER_MDAT && gst_adapter_available (demux->adapter) <= demux->neededbytes) { gst_qtdemux_post_progress (demux, gst_adapter_available (demux->adapter), demux->neededbytes); } done: gst_object_unref (demux); return ret; /* ERRORS */ unknown_stream: { GST_ELEMENT_ERROR (demux, STREAM, FAILED, (NULL), ("unknown stream found")); ret = GST_FLOW_ERROR; goto done; } eos: { GST_DEBUG_OBJECT (demux, "no next entry, EOS"); ret = GST_FLOW_UNEXPECTED; goto done; } invalid_state: { GST_ELEMENT_ERROR (demux, STREAM, FAILED, (NULL), ("qtdemuxer invalid state %d", demux->state)); ret = GST_FLOW_ERROR; goto done; } no_moov: { GST_ELEMENT_ERROR (demux, STREAM, FAILED, (NULL), ("no 'moov' atom withing first 10 MB")); ret = GST_FLOW_ERROR; goto done; } } static gboolean qtdemux_sink_activate (GstPad * sinkpad) { if (gst_pad_check_pull_range (sinkpad)) return gst_pad_activate_pull (sinkpad, TRUE); else return gst_pad_activate_push (sinkpad, TRUE); } static gboolean qtdemux_sink_activate_pull (GstPad * sinkpad, gboolean active) { GstQTDemux *demux = GST_QTDEMUX (GST_PAD_PARENT (sinkpad)); if (active) { demux->pullbased = TRUE; demux->segment_running = TRUE; return gst_pad_start_task (sinkpad, (GstTaskFunction) gst_qtdemux_loop, sinkpad); } else { demux->segment_running = FALSE; return gst_pad_stop_task (sinkpad); } } static gboolean qtdemux_sink_activate_push (GstPad * sinkpad, gboolean active) { GstQTDemux *demux = GST_QTDEMUX (GST_PAD_PARENT (sinkpad)); demux->pullbased = FALSE; return TRUE; } #ifdef HAVE_ZLIB static void * qtdemux_zalloc (void *opaque, unsigned int items, unsigned int size) { return g_malloc (items * size); } static void qtdemux_zfree (void *opaque, void *addr) { g_free (addr); } static void * qtdemux_inflate (void *z_buffer, guint z_length, guint length) { guint8 *buffer; z_stream *z; int ret; z = g_new0 (z_stream, 1); z->zalloc = qtdemux_zalloc; z->zfree = qtdemux_zfree; z->opaque = NULL; z->next_in = z_buffer; z->avail_in = z_length; buffer = (guint8 *) g_malloc (length); ret = inflateInit (z); while (z->avail_in > 0) { if (z->avail_out == 0) { length += 1024; buffer = (guint8 *) g_realloc (buffer, length); z->next_out = buffer + z->total_out; z->avail_out = 1024; } ret = inflate (z, Z_SYNC_FLUSH); if (ret != Z_OK) break; } if (ret != Z_STREAM_END) { g_warning ("inflate() returned %d", ret); } g_free (z); return buffer; } #endif /* HAVE_ZLIB */ static gboolean qtdemux_parse_moov (GstQTDemux * qtdemux, const guint8 * buffer, guint length) { GNode *cmov; qtdemux->moov_node = g_node_new ((guint8 *) buffer); GST_DEBUG_OBJECT (qtdemux, "parsing 'moov' atom"); qtdemux_parse_node (qtdemux, qtdemux->moov_node, buffer, length); cmov = qtdemux_tree_get_child_by_type (qtdemux->moov_node, FOURCC_cmov); if (cmov) { guint32 method; GNode *dcom; GNode *cmvd; dcom = qtdemux_tree_get_child_by_type (cmov, FOURCC_dcom); cmvd = qtdemux_tree_get_child_by_type (cmov, FOURCC_cmvd); if (dcom == NULL || cmvd == NULL) goto invalid_compression; method = QT_FOURCC ((guint8 *) dcom->data + 8); switch (method) { #ifdef HAVE_ZLIB case GST_MAKE_FOURCC ('z', 'l', 'i', 'b'):{ guint uncompressed_length; guint compressed_length; guint8 *buf; uncompressed_length = QT_UINT32 ((guint8 *) cmvd->data + 8); compressed_length = QT_UINT32 ((guint8 *) cmvd->data + 4) - 12; GST_LOG ("length = %u", uncompressed_length); buf = (guint8 *) qtdemux_inflate ((guint8 *) cmvd->data + 12, compressed_length, uncompressed_length); qtdemux->moov_node_compressed = qtdemux->moov_node; qtdemux->moov_node = g_node_new (buf); qtdemux_parse_node (qtdemux, qtdemux->moov_node, buf, uncompressed_length); break; } #endif /* HAVE_ZLIB */ default: GST_WARNING_OBJECT (qtdemux, "unknown or unhandled header compression " "type %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (method)); break; } } return TRUE; /* ERRORS */ invalid_compression: { GST_ERROR_OBJECT (qtdemux, "invalid compressed header"); return FALSE; } } static gboolean qtdemux_parse_container (GstQTDemux * qtdemux, GNode * node, const guint8 * buf, const guint8 * end) { while (G_UNLIKELY (buf < end)) { GNode *child; guint32 len; if (G_UNLIKELY (buf + 4 > end)) { GST_LOG_OBJECT (qtdemux, "buffer overrun"); break; } len = QT_UINT32 (buf); if (G_UNLIKELY (len == 0)) { GST_LOG_OBJECT (qtdemux, "empty container"); break; } if (G_UNLIKELY (len < 8)) { GST_WARNING_OBJECT (qtdemux, "length too short (%d < 8)", len); break; } if (G_UNLIKELY (len > (end - buf))) { GST_WARNING_OBJECT (qtdemux, "length too long (%d > %d)", len, (gint) (end - buf)); break; } child = g_node_new ((guint8 *) buf); g_node_append (node, child); qtdemux_parse_node (qtdemux, child, buf, len); buf += len; } return TRUE; } static gboolean qtdemux_parse_theora_extension (GstQTDemux * qtdemux, QtDemuxStream * stream, GNode * xdxt) { int len = QT_UINT32 (xdxt->data); guint8 *buf = xdxt->data; guint8 *end = buf + len; GstBuffer *buffer; /* skip size and type */ buf += 8; end -= 8; while (buf < end) { gint size; guint32 type; size = QT_UINT32 (buf); type = QT_FOURCC (buf + 4); GST_LOG_OBJECT (qtdemux, "%p %p", buf, end); if (buf + size > end || size <= 0) break; buf += 8; size -= 8; GST_WARNING_OBJECT (qtdemux, "have cookie %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (type)); switch (type) { case FOURCC_tCtH: buffer = gst_buffer_new_and_alloc (size); memcpy (GST_BUFFER_DATA (buffer), buf, size); stream->buffers = g_slist_append (stream->buffers, buffer); GST_LOG_OBJECT (qtdemux, "parsing theora header"); break; case FOURCC_tCt_: buffer = gst_buffer_new_and_alloc (size); memcpy (GST_BUFFER_DATA (buffer), buf, size); stream->buffers = g_slist_append (stream->buffers, buffer); GST_LOG_OBJECT (qtdemux, "parsing theora comment"); break; case FOURCC_tCtC: buffer = gst_buffer_new_and_alloc (size); memcpy (GST_BUFFER_DATA (buffer), buf, size); stream->buffers = g_slist_append (stream->buffers, buffer); GST_LOG_OBJECT (qtdemux, "parsing theora codebook"); break; default: GST_WARNING_OBJECT (qtdemux, "unknown theora cookie %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (type)); break; } buf += size; } return TRUE; } static gboolean qtdemux_parse_node (GstQTDemux * qtdemux, GNode * node, const guint8 * buffer, guint length) { guint32 fourcc = 0; guint32 node_length = 0; const QtNodeType *type; const guint8 *end; GST_LOG_OBJECT (qtdemux, "qtdemux_parse buffer %p length %u", buffer, length); if (G_UNLIKELY (length < 8)) goto not_enough_data; node_length = QT_UINT32 (buffer); fourcc = QT_FOURCC (buffer + 4); /* ignore empty nodes */ if (G_UNLIKELY (fourcc == 0 || node_length == 8)) return TRUE; type = qtdemux_type_get (fourcc); end = buffer + length; GST_LOG_OBJECT (qtdemux, "parsing '%" GST_FOURCC_FORMAT "', length=%u, name '%s'", GST_FOURCC_ARGS (fourcc), node_length, type->name); if (node_length > length) goto broken_atom_size; if (type->flags & QT_FLAG_CONTAINER) { qtdemux_parse_container (qtdemux, node, buffer + 8, end); } else { switch (fourcc) { case FOURCC_stsd: { if (node_length < 20) { GST_LOG_OBJECT (qtdemux, "skipping small stsd box"); break; } GST_DEBUG_OBJECT (qtdemux, "parsing stsd (sample table, sample description) atom"); qtdemux_parse_container (qtdemux, node, buffer + 16, end); break; } case FOURCC_mp4a: { guint32 version; guint32 offset; if (length < 20) { /* small boxes are also inside wave inside the mp4a box */ GST_LOG_OBJECT (qtdemux, "skipping small mp4a box"); break; } version = QT_UINT32 (buffer + 16); GST_DEBUG_OBJECT (qtdemux, "mp4a version 0x%08x", version); /* parse any esds descriptors */ switch (version) { case 0x00000000: offset = 0x24; break; case 0x00010000: offset = 0x34; break; case 0x00020000: offset = 0x58; break; default: GST_WARNING_OBJECT (qtdemux, "unhandled mp4a version 0x%08x", version); offset = 0; break; } if (offset) qtdemux_parse_container (qtdemux, node, buffer + offset, end); break; } case FOURCC_mp4v: { const guint8 *buf; guint32 version; int tlen; GST_DEBUG_OBJECT (qtdemux, "parsing in mp4v"); version = QT_UINT32 (buffer + 16); GST_DEBUG_OBJECT (qtdemux, "version %08x", version); if (1 || version == 0x00000000) { buf = buffer + 0x32; /* FIXME Quicktime uses PASCAL string while * the iso format uses C strings. Check the file * type before attempting to parse the string here. */ tlen = QT_UINT8 (buf); GST_DEBUG_OBJECT (qtdemux, "tlen = %d", tlen); buf++; GST_DEBUG_OBJECT (qtdemux, "string = %.*s", tlen, (char *) buf); /* the string has a reserved space of 32 bytes so skip * the remaining 31 */ buf += 31; buf += 4; /* and 4 bytes reserved */ GST_MEMDUMP_OBJECT (qtdemux, "mp4v", buf, end - buf); qtdemux_parse_container (qtdemux, node, buf, end); } break; } case FOURCC_avc1: { GST_MEMDUMP_OBJECT (qtdemux, "avc1", buffer, end - buffer); qtdemux_parse_container (qtdemux, node, buffer + 0x56, end); break; } case FOURCC_mjp2: { qtdemux_parse_container (qtdemux, node, buffer + 86, end); break; } case FOURCC_meta: { GST_DEBUG_OBJECT (qtdemux, "parsing meta atom"); qtdemux_parse_container (qtdemux, node, buffer + 12, end); break; } case FOURCC_XiTh: { guint32 version; guint32 offset; version = QT_UINT32 (buffer + 12); GST_DEBUG_OBJECT (qtdemux, "parsing XiTh atom version 0x%08x", version); switch (version) { case 0x00000001: offset = 0x62; break; default: GST_DEBUG_OBJECT (qtdemux, "unknown version 0x%08x", version); offset = 0; break; } if (offset) qtdemux_parse_container (qtdemux, node, buffer + offset, end); break; } case FOURCC_in24: { qtdemux_parse_container (qtdemux, node, buffer + 0x34, end); break; } default: if (!strcmp (type->name, "unknown")) GST_MEMDUMP ("Unknown tag", buffer + 4, end - buffer - 4); break; } } GST_LOG_OBJECT (qtdemux, "parsed '%" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (fourcc)); return TRUE; /* ERRORS */ not_enough_data: { GST_ELEMENT_ERROR (qtdemux, STREAM, DEMUX, (_("This file is corrupt and cannot be played.")), ("Not enough data for an atom header, got only %u bytes", length)); return FALSE; } broken_atom_size: { GST_ELEMENT_ERROR (qtdemux, STREAM, DEMUX, (_("This file is corrupt and cannot be played.")), ("Atom '%" GST_FOURCC_FORMAT "' has size of %u bytes, but we have only " "%u bytes available.", GST_FOURCC_ARGS (fourcc), node_length, length)); return FALSE; } } static GNode * qtdemux_tree_get_child_by_type (GNode * node, guint32 fourcc) { GNode *child; guint8 *buffer; guint32 child_fourcc; for (child = g_node_first_child (node); child; child = g_node_next_sibling (child)) { buffer = (guint8 *) child->data; child_fourcc = QT_FOURCC (buffer + 4); if (G_UNLIKELY (child_fourcc == fourcc)) { return child; } } return NULL; } static GNode * qtdemux_tree_get_child_by_type_full (GNode * node, guint32 fourcc, QtAtomParser * parser) { GNode *child; guint8 *buffer; guint32 child_fourcc, child_len; for (child = g_node_first_child (node); child; child = g_node_next_sibling (child)) { buffer = (guint8 *) child->data; child_len = QT_UINT32 (buffer); child_fourcc = QT_FOURCC (buffer + 4); if (G_UNLIKELY (child_fourcc == fourcc)) { if (G_UNLIKELY (child_len < (4 + 4))) return NULL; /* FIXME: must verify if atom length < parent atom length */ qt_atom_parser_init (parser, buffer + (4 + 4), child_len - (4 + 4)); return child; } } return NULL; } static GNode * qtdemux_tree_get_sibling_by_type (GNode * node, guint32 fourcc) { GNode *child; guint8 *buffer; guint32 child_fourcc; for (child = g_node_next_sibling (node); child; child = g_node_next_sibling (child)) { buffer = (guint8 *) child->data; child_fourcc = QT_FOURCC (buffer + 4); if (child_fourcc == fourcc) { return child; } } return NULL; } static gboolean gst_qtdemux_add_stream (GstQTDemux * qtdemux, QtDemuxStream * stream, GstTagList * list) { if (qtdemux->n_streams >= GST_QTDEMUX_MAX_STREAMS) goto too_many_streams; if (stream->subtype == FOURCC_vide) { gchar *name = g_strdup_printf ("video_%02d", qtdemux->n_video_streams); stream->pad = gst_pad_new_from_static_template (&gst_qtdemux_videosrc_template, name); g_free (name); /* fps is calculated base on the duration of the first frames since * qt does not have a fixed framerate. */ if ((stream->n_samples == 1) && (stream->min_duration == 0)) { /* still frame */ stream->fps_n = 0; stream->fps_d = 1; } else { stream->fps_n = stream->timescale; if (stream->min_duration == 0) stream->fps_d = 1; else stream->fps_d = stream->min_duration; } if (stream->caps) { gboolean gray; gint depth, palette_count; const guint32 *palette_data = NULL; gst_caps_set_simple (stream->caps, "width", G_TYPE_INT, stream->width, "height", G_TYPE_INT, stream->height, "framerate", GST_TYPE_FRACTION, stream->fps_n, stream->fps_d, NULL); /* iso files: * calculate pixel-aspect-ratio using display width and height */ if (qtdemux->major_brand != FOURCC_qt__) { GST_DEBUG_OBJECT (qtdemux, "video size %dx%d, target display size %dx%d", stream->width, stream->height, stream->display_width, stream->display_height); if (stream->display_width > 0 && stream->display_height > 0 && stream->width > 0 && stream->height > 0) { gint n, d; /* calculate the pixel aspect ratio using the display and pixel w/h */ n = stream->display_width * stream->height; d = stream->display_height * stream->width; if (n != d) { GST_DEBUG_OBJECT (qtdemux, "setting PAR to %d/%d", n, d); gst_caps_set_simple (stream->caps, "pixel-aspect-ratio", GST_TYPE_FRACTION, n, d, NULL); } } } /* qt file might have pasp atom */ if (stream->par_w > 0 && stream->par_h > 0) { GST_DEBUG_OBJECT (qtdemux, "par %d:%d", stream->par_w, stream->par_h); gst_caps_set_simple (stream->caps, "pixel-aspect-ratio", GST_TYPE_FRACTION, stream->par_w, stream->par_h, NULL); } depth = stream->bits_per_sample; /* more than 32 bits means grayscale */ gray = (depth > 32); /* low 32 bits specify the depth */ depth &= 0x1F; /* different number of palette entries is determined by depth. */ palette_count = 0; if ((depth == 1) || (depth == 2) || (depth == 4) || (depth == 8)) palette_count = (1 << depth); switch (palette_count) { case 0: break; case 2: palette_data = ff_qt_default_palette_2; break; case 4: palette_data = ff_qt_default_palette_4; break; case 16: if (gray) palette_data = ff_qt_grayscale_palette_16; else palette_data = ff_qt_default_palette_16; break; case 256: if (gray) palette_data = ff_qt_grayscale_palette_256; else palette_data = ff_qt_default_palette_256; break; default: GST_ELEMENT_WARNING (qtdemux, STREAM, DECODE, (_("The video in this file might not play correctly.")), ("unsupported palette depth %d", depth)); break; } if (palette_data) { GstBuffer *palette; /* make sure it's not writable. We leave MALLOCDATA to NULL so that we * don't free any of the buffer data. */ palette = gst_buffer_new (); GST_BUFFER_FLAG_SET (palette, GST_BUFFER_FLAG_READONLY); GST_BUFFER_DATA (palette) = (guint8 *) palette_data; GST_BUFFER_SIZE (palette) = sizeof (guint32) * palette_count; gst_caps_set_simple (stream->caps, "palette_data", GST_TYPE_BUFFER, palette, NULL); gst_buffer_unref (palette); } else if (palette_count != 0) { GST_ELEMENT_WARNING (qtdemux, STREAM, NOT_IMPLEMENTED, (NULL), ("Unsupported palette depth %d. Ignoring stream.", depth)); gst_object_unref (stream->pad); stream->pad = NULL; } } qtdemux->n_video_streams++; } else if (stream->subtype == FOURCC_soun) { gchar *name = g_strdup_printf ("audio_%02d", qtdemux->n_audio_streams); stream->pad = gst_pad_new_from_static_template (&gst_qtdemux_audiosrc_template, name); g_free (name); if (stream->caps) { gst_caps_set_simple (stream->caps, "rate", G_TYPE_INT, (int) stream->rate, "channels", G_TYPE_INT, stream->n_channels, NULL); } qtdemux->n_audio_streams++; } else if (stream->subtype == FOURCC_strm) { GST_DEBUG_OBJECT (qtdemux, "stream type, not creating pad"); } else if (stream->subtype == FOURCC_subp || stream->subtype == FOURCC_text) { gchar *name = g_strdup_printf ("subtitle_%02d", qtdemux->n_sub_streams); stream->pad = gst_pad_new_from_static_template (&gst_qtdemux_subsrc_template, name); g_free (name); qtdemux->n_sub_streams++; } else { GST_DEBUG_OBJECT (qtdemux, "unknown stream type"); goto done; } qtdemux->streams[qtdemux->n_streams] = stream; qtdemux->n_streams++; GST_DEBUG_OBJECT (qtdemux, "n_streams is now %d", qtdemux->n_streams); if (stream->pad) { GST_PAD_ELEMENT_PRIVATE (stream->pad) = stream; gst_pad_use_fixed_caps (stream->pad); gst_pad_set_event_function (stream->pad, gst_qtdemux_handle_src_event); gst_pad_set_query_type_function (stream->pad, gst_qtdemux_get_src_query_types); gst_pad_set_query_function (stream->pad, gst_qtdemux_handle_src_query); GST_DEBUG_OBJECT (qtdemux, "setting caps %" GST_PTR_FORMAT, stream->caps); gst_pad_set_caps (stream->pad, stream->caps); GST_DEBUG_OBJECT (qtdemux, "adding pad %s %p to qtdemux %p", GST_OBJECT_NAME (stream->pad), stream->pad, qtdemux); gst_pad_set_active (stream->pad, TRUE); gst_element_add_pad (GST_ELEMENT_CAST (qtdemux), stream->pad); if (stream->pending_tags) gst_tag_list_free (stream->pending_tags); stream->pending_tags = list; if (list) { /* post now, send event on pad later */ GST_DEBUG_OBJECT (qtdemux, "Posting tags %" GST_PTR_FORMAT, list); gst_element_post_message (GST_ELEMENT (qtdemux), gst_message_new_tag_full (GST_OBJECT (qtdemux), stream->pad, gst_tag_list_copy (list))); } /* global tags go on each pad anyway */ stream->send_global_tags = TRUE; } done: return TRUE; too_many_streams: { GST_ELEMENT_WARNING (qtdemux, STREAM, DECODE, (_("This file contains too many streams. Only playing first %d"), GST_QTDEMUX_MAX_STREAMS), (NULL)); return TRUE; } } /* collect all samples for @stream by reading the info from @stbl */ static gboolean qtdemux_parse_samples (GstQTDemux * qtdemux, QtDemuxStream * stream, GNode * stbl) { QtAtomParser co_reader; QtAtomParser stsz; QtAtomParser stsc; QtAtomParser stts; GNode *ctts; guint32 sample_size; guint32 n_samples; guint32 n_samples_per_chunk; int sample_index; QtDemuxSample *samples; gint i, j, k; int index; guint64 timestamp; guint co_size; /* sample to chunk */ if (!qtdemux_tree_get_child_by_type_full (stbl, FOURCC_stsc, &stsc)) goto corrupt_file; /* sample size */ if (!qtdemux_tree_get_child_by_type_full (stbl, FOURCC_stsz, &stsz)) goto corrupt_file; /* chunk offsets */ if (qtdemux_tree_get_child_by_type_full (stbl, FOURCC_stco, &co_reader)) co_size = sizeof (guint32); else if (qtdemux_tree_get_child_by_type_full (stbl, FOURCC_co64, &co_reader)) co_size = sizeof (guint64); else goto corrupt_file; /* sample time */ if (!qtdemux_tree_get_child_by_type_full (stbl, FOURCC_stts, &stts)) goto corrupt_file; if (!qt_atom_parser_skip (&stsz, 1 + 3) || !qt_atom_parser_get_uint32 (&stsz, &sample_size)) goto corrupt_file; if (sample_size == 0 || stream->sampled) { /* skip version, flags, number of entries */ if (!gst_byte_reader_skip (&co_reader, 1 + 3 + 4)) goto corrupt_file; if (!qt_atom_parser_get_uint32 (&stsz, &n_samples)) goto corrupt_file; if (n_samples == 0) goto no_samples; GST_DEBUG_OBJECT (qtdemux, "stsz sample_size 0, allocating n_samples %u " "(%u MB)", n_samples, (guint) (n_samples * sizeof (QtDemuxSample)) >> 20); if (n_samples >= QTDEMUX_MAX_SAMPLE_INDEX_SIZE / sizeof (QtDemuxSample)) goto index_too_big; samples = g_try_new0 (QtDemuxSample, n_samples); if (samples == NULL) goto out_of_memory; stream->n_samples = n_samples; stream->samples = samples; /* set the sample sizes */ if (sample_size == 0) { /* different sizes for each sample */ if (!qt_atom_parser_has_chunks (&stsz, n_samples, 4)) goto corrupt_file; for (i = 0; i < n_samples; i++) { samples[i].size = qt_atom_parser_get_uint32_unchecked (&stsz); GST_LOG_OBJECT (qtdemux, "sample %d has size %u", i, samples[i].size); } } else { /* samples have the same size */ GST_LOG_OBJECT (qtdemux, "all samples have size %u", sample_size); for (i = 0; i < n_samples; i++) samples[i].size = sample_size; } /* set the sample offsets in the file */ if (!qt_atom_parser_skip (&stsc, 1 + 3) || !qt_atom_parser_get_uint32 (&stsc, &n_samples_per_chunk)) goto corrupt_file; if (!qt_atom_parser_has_chunks (&stsc, n_samples_per_chunk, 12)) goto corrupt_file; index = 0; for (i = 0; i < n_samples_per_chunk; i++) { QtAtomParser co_chunk; guint32 first_chunk, last_chunk; guint32 samples_per_chunk; first_chunk = qt_atom_parser_get_uint32_unchecked (&stsc); samples_per_chunk = qt_atom_parser_get_uint32_unchecked (&stsc); qt_atom_parser_skip_unchecked (&stsc, 4); /* chunk numbers are counted from 1 it seems */ if (G_UNLIKELY (first_chunk == 0)) goto corrupt_file; else --first_chunk; /* the last chunk of each entry is calculated by taking the first chunk * of the next entry; except if there is no next, where we fake it with * INT_MAX */ if (G_UNLIKELY (i == n_samples_per_chunk - 1)) { last_chunk = G_MAXUINT32; } else { last_chunk = qt_atom_parser_peek_uint32_unchecked (&stsc); if (G_UNLIKELY (last_chunk == 0)) goto corrupt_file; else --last_chunk; } if (G_UNLIKELY (last_chunk < first_chunk)) goto corrupt_file; if (last_chunk != G_MAXUINT32) { if (!qt_atom_parser_peek_sub (&co_reader, first_chunk * co_size, (last_chunk - first_chunk) * co_size, &co_chunk)) goto corrupt_file; } else { co_chunk = co_reader; if (!qt_atom_parser_skip (&co_chunk, first_chunk * co_size)) goto corrupt_file; } for (j = first_chunk; j < last_chunk; j++) { guint64 chunk_offset; if (!qt_atom_parser_get_offset (&co_chunk, co_size, &chunk_offset)) goto corrupt_file; for (k = 0; k < samples_per_chunk; k++) { GST_LOG_OBJECT (qtdemux, "Creating entry %d with offset %" G_GUINT64_FORMAT, index, chunk_offset); samples[index].offset = chunk_offset; chunk_offset += samples[index].size; index++; if (G_UNLIKELY (index >= n_samples)) goto done2; } } } done2: { guint32 n_sample_times; guint32 time; if (!qt_atom_parser_skip (&stts, 4)) goto corrupt_file; if (!qt_atom_parser_get_uint32 (&stts, &n_sample_times)) goto corrupt_file; GST_LOG_OBJECT (qtdemux, "%u timestamp blocks", n_sample_times); /* make sure there's enough data */ if (!qt_atom_parser_has_chunks (&stts, n_sample_times, 2 * 4)) goto corrupt_file; timestamp = 0; stream->min_duration = 0; time = 0; index = 0; for (i = 0; i < n_sample_times; i++) { guint32 n; guint32 duration; n = qt_atom_parser_get_uint32_unchecked (&stts); duration = qt_atom_parser_get_uint32_unchecked (&stts); GST_LOG_OBJECT (qtdemux, "block %d, %u timestamps, duration %u ", i, n, duration); /* take first duration for fps */ if (G_UNLIKELY (stream->min_duration == 0)) stream->min_duration = duration; for (j = 0; j < n; j++) { GST_DEBUG_OBJECT (qtdemux, "sample %d: index %d, timestamp %" GST_TIME_FORMAT, index, j, GST_TIME_ARGS (timestamp)); samples[index].timestamp = timestamp; /* add non-scaled values to avoid rounding errors */ time += duration; timestamp = gst_util_uint64_scale (time, GST_SECOND, stream->timescale); samples[index].duration = timestamp - samples[index].timestamp; index++; if (G_UNLIKELY (index >= n_samples)) goto done3; } } /* fill up empty timestamps with the last timestamp, this can happen when * the last samples do not decode and so we don't have timestamps for them. * We however look at the last timestamp to estimate the track length so we * need something in here. */ for (; index < n_samples; index++) { GST_DEBUG_OBJECT (qtdemux, "fill sample %d: timestamp %" GST_TIME_FORMAT, index, GST_TIME_ARGS (timestamp)); samples[index].timestamp = timestamp; samples[index].duration = -1; } } done3: { /* FIXME: split this block out into a separate function */ QtAtomParser stss, stps; /* sample sync, can be NULL */ if (qtdemux_tree_get_child_by_type_full (stbl, FOURCC_stss, &stss)) { guint32 n_sample_syncs; /* mark keyframes */ if (!qt_atom_parser_skip (&stss, 4)) goto corrupt_file; if (!qt_atom_parser_get_uint32 (&stss, &n_sample_syncs)) goto corrupt_file; if (n_sample_syncs == 0) { stream->all_keyframe = TRUE; } else { /* make sure there's enough data */ if (!qt_atom_parser_has_chunks (&stss, n_sample_syncs, 4)) goto corrupt_file; for (i = 0; i < n_sample_syncs; i++) { /* note that the first sample is index 1, not 0 */ index = qt_atom_parser_get_uint32_unchecked (&stss); if (G_LIKELY (index > 0 && index <= n_samples)) samples[index - 1].keyframe = TRUE; } } /* stps marks partial sync frames like open GOP I-Frames */ if (qtdemux_tree_get_child_by_type_full (stbl, FOURCC_stps, &stps)) { guint32 n_sample_syncs; if (!qt_atom_parser_skip (&stps, 4)) goto corrupt_file; if (!qt_atom_parser_get_uint32 (&stps, &n_sample_syncs)) goto corrupt_file; if (n_sample_syncs != 0) { /* no entries, the stss table contains the real sync * samples */ } else { /* make sure there's enough data */ if (!qt_atom_parser_has_chunks (&stps, n_sample_syncs, 4)) goto corrupt_file; for (i = 0; i < n_sample_syncs; i++) { /* note that the first sample is index 1, not 0 */ index = qt_atom_parser_get_uint32_unchecked (&stps); if (G_LIKELY (index > 0 && index <= n_samples)) samples[index - 1].keyframe = TRUE; } } } } else { /* no stss, all samples are keyframes */ stream->all_keyframe = TRUE; } } } else { GST_DEBUG_OBJECT (qtdemux, "stsz sample_size %d != 0, treating chunks as samples", sample_size); /* skip version + flags */ if (!gst_byte_reader_skip (&co_reader, 1 + 3)) goto corrupt_file; /* treat chunks as samples */ if (!gst_byte_reader_get_uint32_be (&co_reader, &n_samples)) goto corrupt_file; if (n_samples == 0) goto no_samples; GST_DEBUG_OBJECT (qtdemux, "allocating n_samples %u (%u MB)", n_samples, (guint) (n_samples * sizeof (QtDemuxSample)) >> 20); if (n_samples >= QTDEMUX_MAX_SAMPLE_INDEX_SIZE / sizeof (QtDemuxSample)) goto index_too_big; samples = g_try_new0 (QtDemuxSample, n_samples); if (samples == NULL) goto out_of_memory; stream->n_samples = n_samples; stream->samples = samples; if (!qt_atom_parser_skip (&stsc, 1 + 3) || !qt_atom_parser_get_uint32 (&stsc, &n_samples_per_chunk)) goto corrupt_file; GST_DEBUG_OBJECT (qtdemux, "n_samples_per_chunk %u", n_samples_per_chunk); sample_index = 0; timestamp = 0; if (!qt_atom_parser_has_chunks (&stsc, n_samples_per_chunk, 12)) goto corrupt_file; for (i = 0; i < n_samples_per_chunk; i++) { QtAtomParser co_chunk; guint32 first_chunk, last_chunk; guint32 samples_per_chunk; first_chunk = qt_atom_parser_get_uint32_unchecked (&stsc); samples_per_chunk = qt_atom_parser_get_uint32_unchecked (&stsc); qt_atom_parser_skip_unchecked (&stsc, 4); /* chunk numbers are counted from 1 it seems */ if (G_UNLIKELY (first_chunk == 0)) goto corrupt_file; else --first_chunk; /* the last chunk of each entry is calculated by taking the first chunk * of the next entry; except if there is no next, where we fake it with * INT_MAX */ if (G_UNLIKELY (i == (n_samples_per_chunk - 1))) { last_chunk = G_MAXUINT32; } else { last_chunk = qt_atom_parser_peek_uint32_unchecked (&stsc); if (G_UNLIKELY (last_chunk == 0)) goto corrupt_file; else --last_chunk; } GST_LOG_OBJECT (qtdemux, "entry %d has first_chunk %d, last_chunk %d, samples_per_chunk %d", i, first_chunk, last_chunk, samples_per_chunk); if (G_UNLIKELY (last_chunk < first_chunk)) goto corrupt_file; if (last_chunk != G_MAXUINT32) { if (!qt_atom_parser_peek_sub (&co_reader, first_chunk * co_size, (last_chunk - first_chunk) * co_size, &co_chunk)) goto corrupt_file; } else { co_chunk = co_reader; if (!qt_atom_parser_skip (&co_chunk, first_chunk * co_size)) goto corrupt_file; } for (j = first_chunk; j < last_chunk; j++) { if (j >= n_samples) goto done; samples[j].offset = qt_atom_parser_get_offset_unchecked (&co_chunk, co_size); GST_LOG_OBJECT (qtdemux, "Created entry %d with offset " "%" G_GUINT64_FORMAT, j, samples[j].offset); if (stream->samples_per_frame * stream->bytes_per_frame) { samples[j].size = (samples_per_chunk * stream->n_channels) / stream->samples_per_frame * stream->bytes_per_frame; } else { samples[j].size = samples_per_chunk; } GST_DEBUG_OBJECT (qtdemux, "sample %d: timestamp %" GST_TIME_FORMAT ", size %u", j, GST_TIME_ARGS (timestamp), samples[j].size); samples[j].timestamp = timestamp; sample_index += samples_per_chunk; timestamp = gst_util_uint64_scale (sample_index, GST_SECOND, stream->timescale); samples[j].duration = timestamp - samples[j].timestamp; samples[j].keyframe = TRUE; } } } /* composition time to sample */ if ((ctts = qtdemux_tree_get_child_by_type (stbl, FOURCC_ctts))) { const guint8 *ctts_data, *ctts_p; guint32 n_entries; guint32 count; gint32 soffset; ctts_data = (const guint8 *) ctts->data; n_entries = QT_UINT32 (ctts_data + 12); /* Fill in the pts_offsets */ index = 0; ctts_p = ctts_data + 16; /* FIXME: make sure we don't read beyond the atom size/boundary */ for (i = 0; i < n_entries; i++) { count = QT_UINT32 (ctts_p); ctts_p += 4; soffset = QT_UINT32 (ctts_p); ctts_p += 4; for (j = 0; j < count; j++) { /* we operate with very small soffset values here, it shouldn't overflow */ samples[index].pts_offset = soffset * GST_SECOND / stream->timescale; index++; if (G_UNLIKELY (index >= n_samples)) goto done; } } } done: return TRUE; /* ERRORS */ corrupt_file: { GST_ELEMENT_ERROR (qtdemux, STREAM, DECODE, (_("This file is corrupt and cannot be played.")), (NULL)); return FALSE; } no_samples: { GST_WARNING_OBJECT (qtdemux, "stream has no samples"); return FALSE; } out_of_memory: { GST_WARNING_OBJECT (qtdemux, "failed to allocate %d samples", n_samples); return FALSE; } index_too_big: { GST_WARNING_OBJECT (qtdemux, "not allocating index of %d samples, would " "be larger than %uMB (broken file?)", n_samples, QTDEMUX_MAX_SAMPLE_INDEX_SIZE >> 20); return FALSE; } } /* collect all segment info for @stream. */ static gboolean qtdemux_parse_segments (GstQTDemux * qtdemux, QtDemuxStream * stream, GNode * trak) { GNode *edts; /* parse and prepare segment info from the edit list */ GST_DEBUG_OBJECT (qtdemux, "looking for edit list container"); stream->n_segments = 0; stream->segments = NULL; if ((edts = qtdemux_tree_get_child_by_type (trak, FOURCC_edts))) { GNode *elst; gint n_segments; gint i, count; guint64 time, stime; guint8 *buffer; GST_DEBUG_OBJECT (qtdemux, "looking for edit list"); if (!(elst = qtdemux_tree_get_child_by_type (edts, FOURCC_elst))) goto done; buffer = elst->data; n_segments = QT_UINT32 (buffer + 12); /* we might allocate a bit too much, at least allocate 1 segment */ stream->segments = g_new (QtDemuxSegment, MAX (n_segments, 1)); /* segments always start from 0 */ time = 0; stime = 0; count = 0; for (i = 0; i < n_segments; i++) { guint64 duration; guint64 media_time; QtDemuxSegment *segment; guint32 rate_int; media_time = QT_UINT32 (buffer + 20 + i * 12); /* -1 media time is an empty segment, just ignore it */ if (media_time == G_MAXUINT32) continue; duration = QT_UINT32 (buffer + 16 + i * 12); segment = &stream->segments[count++]; /* time and duration expressed in global timescale */ segment->time = stime; /* add non scaled values so we don't cause roundoff errors */ time += duration; stime = gst_util_uint64_scale (time, GST_SECOND, qtdemux->timescale); segment->stop_time = stime; segment->duration = stime - segment->time; /* media_time expressed in stream timescale */ segment->media_start = gst_util_uint64_scale (media_time, GST_SECOND, stream->timescale); segment->media_stop = segment->media_start + segment->duration; rate_int = GST_READ_UINT32_BE (buffer + 24 + i * 12); if (rate_int <= 1) { /* 0 is not allowed, some programs write 1 instead of the floating point * value */ GST_WARNING_OBJECT (qtdemux, "found suspicious rate %" G_GUINT32_FORMAT, rate_int); segment->rate = 1; } else { segment->rate = rate_int / 65536.0; } GST_DEBUG_OBJECT (qtdemux, "created segment %d time %" GST_TIME_FORMAT ", duration %" GST_TIME_FORMAT ", media_time %" GST_TIME_FORMAT ", rate %g, (%d)", i, GST_TIME_ARGS (segment->time), GST_TIME_ARGS (segment->duration), GST_TIME_ARGS (segment->media_start), segment->rate, rate_int); } GST_DEBUG_OBJECT (qtdemux, "found %d non-empty segments", count); stream->n_segments = count; } done: /* push based does not handle segments, so act accordingly here, * and warn if applicable */ if (!qtdemux->pullbased) { GST_WARNING_OBJECT (qtdemux, "streaming; discarding edit list segments"); /* remove and use default one below, we stream like it anyway */ g_free (stream->segments); stream->segments = NULL; stream->n_segments = 0; } /* no segments, create one to play the complete trak */ if (stream->n_segments == 0) { GstClockTime stream_duration = 0; if (stream->segments == NULL) stream->segments = g_new (QtDemuxSegment, 1); /* samples know best */ if (stream->n_samples > 0) { stream_duration = stream->samples[stream->n_samples - 1].timestamp + stream->samples[stream->n_samples - 1].pts_offset + stream->samples[stream->n_samples - 1].duration; } stream->segments[0].time = 0; stream->segments[0].stop_time = stream_duration; stream->segments[0].duration = stream_duration; stream->segments[0].media_start = 0; stream->segments[0].media_stop = stream_duration; stream->segments[0].rate = 1.0; GST_DEBUG_OBJECT (qtdemux, "created dummy segment %" GST_TIME_FORMAT, GST_TIME_ARGS (stream_duration)); stream->n_segments = 1; } GST_DEBUG_OBJECT (qtdemux, "using %d segments", stream->n_segments); return TRUE; } /* parse the traks. * With each track we associate a new QtDemuxStream that contains all the info * about the trak. * traks that do not decode to something (like strm traks) will not have a pad. */ static gboolean qtdemux_parse_trak (GstQTDemux * qtdemux, GNode * trak) { QtAtomParser tkhd; int offset; GNode *mdia; GNode *mdhd; GNode *hdlr; GNode *minf; GNode *stbl; GNode *stsd; GNode *mp4a; GNode *mp4v; GNode *wave; GNode *esds; GNode *pasp; QtDemuxStream *stream; GstTagList *list = NULL; gchar *codec = NULL; const guint8 *stsd_data; guint32 version; guint32 tkhd_flags; guint8 tkhd_version; guint32 fourcc; guint len; stream = g_new0 (QtDemuxStream, 1); /* new streams always need a discont */ stream->discont = TRUE; /* we enable clipping for raw audio/video streams */ stream->need_clip = FALSE; stream->need_process = FALSE; stream->segment_index = -1; stream->time_position = 0; stream->sample_index = -1; stream->last_ret = GST_FLOW_OK; if (!qtdemux_tree_get_child_by_type_full (trak, FOURCC_tkhd, &tkhd) || !qt_atom_parser_get_uint8 (&tkhd, &tkhd_version) || !qt_atom_parser_get_uint24 (&tkhd, &tkhd_flags)) goto corrupt_file; GST_LOG_OBJECT (qtdemux, "track[tkhd] version/flags: 0x%02x/%06x", tkhd_version, tkhd_flags); if (!(mdia = qtdemux_tree_get_child_by_type (trak, FOURCC_mdia))) goto corrupt_file; if (!(mdhd = qtdemux_tree_get_child_by_type (mdia, FOURCC_mdhd))) { /* be nice for some crooked mjp2 files that use mhdr for mdhd */ if (qtdemux->major_brand != FOURCC_mjp2 || !(mdhd = qtdemux_tree_get_child_by_type (mdia, FOURCC_mhdr))) goto corrupt_file; } version = QT_UINT32 ((guint8 *) mdhd->data + 8); GST_LOG_OBJECT (qtdemux, "track version/flags: %08x", version); if (version == 0x01000000) { stream->timescale = QT_UINT32 ((guint8 *) mdhd->data + 28); stream->duration = QT_UINT64 ((guint8 *) mdhd->data + 32); } else { stream->timescale = QT_UINT32 ((guint8 *) mdhd->data + 20); stream->duration = QT_UINT32 ((guint8 *) mdhd->data + 24); } GST_LOG_OBJECT (qtdemux, "track timescale: %" G_GUINT32_FORMAT, stream->timescale); GST_LOG_OBJECT (qtdemux, "track duration: %" G_GUINT64_FORMAT, stream->duration); if (G_UNLIKELY (stream->timescale == 0 || qtdemux->timescale == 0)) goto corrupt_file; if (qtdemux->duration != G_MAXINT32 && stream->duration != G_MAXINT32) { guint64 tdur1, tdur2; /* don't overflow */ tdur1 = stream->timescale * (guint64) qtdemux->duration; tdur2 = qtdemux->timescale * (guint64) stream->duration; /* HACK: * some of those trailers, nowadays, have prologue images that are * themselves vide tracks as well. I haven't really found a way to * identify those yet, except for just looking at their duration. */ if (tdur1 != 0 && (tdur2 * 10 / tdur1) < 2) { GST_WARNING_OBJECT (qtdemux, "Track shorter than 20%% (%" G_GUINT64_FORMAT "/%" G_GUINT32_FORMAT " vs. %" G_GUINT32_FORMAT "/%" G_GUINT32_FORMAT ") of the stream " "found, assuming preview image or something; skipping track", stream->duration, stream->timescale, qtdemux->duration, qtdemux->timescale); g_free (stream); return TRUE; } } if (!(hdlr = qtdemux_tree_get_child_by_type (mdia, FOURCC_hdlr))) goto corrupt_file; GST_LOG_OBJECT (qtdemux, "track type: %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (QT_FOURCC ((guint8 *) hdlr->data + 12))); stream->subtype = QT_FOURCC ((guint8 *) hdlr->data + 16); GST_LOG_OBJECT (qtdemux, "track subtype: %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (stream->subtype)); if (!(minf = qtdemux_tree_get_child_by_type (mdia, FOURCC_minf))) goto corrupt_file; if (!(stbl = qtdemux_tree_get_child_by_type (minf, FOURCC_stbl))) goto corrupt_file; /* parse stsd */ if (!(stsd = qtdemux_tree_get_child_by_type (stbl, FOURCC_stsd))) goto corrupt_file; stsd_data = (const guint8 *) stsd->data; /* stsd should at least have one entry */ len = QT_UINT32 (stsd_data); if (len < 24) goto corrupt_file; /* and that entry should fit within stsd */ len = QT_UINT32 (stsd_data + 16); if (len > QT_UINT32 (stsd_data) + 16) goto corrupt_file; GST_LOG_OBJECT (qtdemux, "stsd len: %d", len); stream->fourcc = fourcc = QT_FOURCC (stsd_data + 16 + 4); GST_LOG_OBJECT (qtdemux, "stsd type: %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (stream->fourcc)); if ((fourcc == FOURCC_drms) || (fourcc == FOURCC_drmi) || ((fourcc & 0xFFFFFF00) == GST_MAKE_FOURCC ('e', 'n', 'c', 0))) goto error_encrypted; if (stream->subtype == FOURCC_vide) { guint32 w, h; stream->sampled = TRUE; /* version 1 uses some 64-bit ints */ if (!qt_atom_parser_skip (&tkhd, (tkhd_version == 1) ? 84 : 72) || !qt_atom_parser_get_uint32 (&tkhd, &w) || !qt_atom_parser_get_uint32 (&tkhd, &h)) goto corrupt_file; stream->display_width = w >> 16; stream->display_height = h >> 16; offset = 16; if (len < 86) goto corrupt_file; stream->width = QT_UINT16 (stsd_data + offset + 32); stream->height = QT_UINT16 (stsd_data + offset + 34); stream->fps_n = 0; /* this is filled in later */ stream->fps_d = 0; /* this is filled in later */ stream->bits_per_sample = QT_UINT16 (stsd_data + offset + 82); stream->color_table_id = QT_UINT16 (stsd_data + offset + 84); GST_LOG_OBJECT (qtdemux, "frame count: %u", QT_UINT16 (stsd_data + offset + 48)); stream->caps = qtdemux_video_caps (qtdemux, stream, fourcc, stsd_data, &codec); if (codec) { list = gst_tag_list_new (); gst_tag_list_add (list, GST_TAG_MERGE_REPLACE, GST_TAG_VIDEO_CODEC, codec, NULL); g_free (codec); codec = NULL; } esds = NULL; pasp = NULL; mp4v = qtdemux_tree_get_child_by_type (stsd, FOURCC_mp4v); /* H264 is MPEG-4 after all, * and qt seems to put MPEG-4 stuff in there as well */ if (!mp4v) mp4v = qtdemux_tree_get_child_by_type (stsd, FOURCC_avc1); if (mp4v) { esds = qtdemux_tree_get_child_by_type (mp4v, FOURCC_esds); pasp = qtdemux_tree_get_child_by_type (mp4v, FOURCC_pasp); } if (pasp) { const guint8 *pasp_data = (const guint8 *) pasp->data; stream->par_w = QT_UINT32 (pasp_data + 8); stream->par_h = QT_UINT32 (pasp_data + 12); } else { stream->par_w = 0; stream->par_h = 0; } if (esds) { gst_qtdemux_handle_esds (qtdemux, stream, esds, list); } else { switch (fourcc) { case FOURCC_avc1: { gint len = QT_UINT32 (stsd_data) - 0x66; const guint8 *avc_data = stsd_data + 0x66; /* find avcC */ while (len >= 0x8 && QT_FOURCC (avc_data + 0x4) != FOURCC_avcC && QT_UINT32 (avc_data) < len) { len -= QT_UINT32 (avc_data); avc_data += QT_UINT32 (avc_data); } /* parse, if found */ if (len > 0x8 && QT_FOURCC (avc_data + 0x4) == FOURCC_avcC) { GstBuffer *buf; gint size; if (QT_UINT32 (avc_data) < len) size = QT_UINT32 (avc_data) - 0x8; else size = len - 0x8; GST_DEBUG_OBJECT (qtdemux, "found avcC codec_data in stsd"); buf = gst_buffer_new_and_alloc (size); memcpy (GST_BUFFER_DATA (buf), avc_data + 0x8, size); gst_caps_set_simple (stream->caps, "codec_data", GST_TYPE_BUFFER, buf, NULL); gst_buffer_unref (buf); } break; } case FOURCC_mjp2: { GNode *jp2h, *colr, *mjp2, *field, *prefix; const guint8 *data; guint32 fourcc = 0; GST_DEBUG_OBJECT (qtdemux, "found mjp2"); /* some required atoms */ mjp2 = qtdemux_tree_get_child_by_type (stsd, FOURCC_mjp2); if (!mjp2) break; jp2h = qtdemux_tree_get_child_by_type (mjp2, FOURCC_jp2h); if (!jp2h) break; colr = qtdemux_tree_get_child_by_type (jp2h, FOURCC_colr); if (!colr) break; GST_DEBUG_OBJECT (qtdemux, "found colr"); /* try to extract colour space info */ if (QT_UINT8 ((guint8 *) colr->data + 8) == 1) { switch (QT_UINT32 ((guint8 *) colr->data + 11)) { case 16: fourcc = GST_MAKE_FOURCC ('s', 'R', 'G', 'B'); break; case 17: fourcc = GST_MAKE_FOURCC ('G', 'R', 'A', 'Y'); break; case 18: fourcc = GST_MAKE_FOURCC ('s', 'Y', 'U', 'V'); break; default: break; } } if (fourcc) gst_caps_set_simple (stream->caps, "fourcc", GST_TYPE_FOURCC, fourcc, NULL); /* some optional atoms */ field = qtdemux_tree_get_child_by_type (mjp2, FOURCC_fiel); prefix = qtdemux_tree_get_child_by_type (mjp2, FOURCC_jp2x); /* indicate possible fields in caps */ if (field) { data = (guint8 *) field->data + 8; if (*data != 1) gst_caps_set_simple (stream->caps, "fields", G_TYPE_INT, (gint) * data, NULL); } /* add codec_data if provided */ if (prefix) { GstBuffer *buf; gint len; GST_DEBUG_OBJECT (qtdemux, "found prefix data in stsd"); data = prefix->data; len = QT_UINT32 (data); if (len > 0x8) { len -= 0x8; buf = gst_buffer_new_and_alloc (len); memcpy (GST_BUFFER_DATA (buf), data + 8, len); gst_caps_set_simple (stream->caps, "codec_data", GST_TYPE_BUFFER, buf, NULL); gst_buffer_unref (buf); } } break; } case FOURCC_SVQ3: case FOURCC_VP31: { GstBuffer *buf; gint len = QT_UINT32 (stsd_data); GST_DEBUG_OBJECT (qtdemux, "found codec_data in stsd"); buf = gst_buffer_new_and_alloc (len); memcpy (GST_BUFFER_DATA (buf), stsd_data, len); gst_caps_set_simple (stream->caps, "codec_data", GST_TYPE_BUFFER, buf, NULL); gst_buffer_unref (buf); break; } case FOURCC_rle_: { gst_caps_set_simple (stream->caps, "depth", G_TYPE_INT, QT_UINT16 (stsd_data + offset + 82), NULL); break; } case FOURCC_XiTh: { GNode *xith, *xdxt; GST_DEBUG_OBJECT (qtdemux, "found XiTh"); xith = qtdemux_tree_get_child_by_type (stsd, FOURCC_XiTh); if (!xith) break; xdxt = qtdemux_tree_get_child_by_type (xith, FOURCC_XdxT); if (!xdxt) break; GST_DEBUG_OBJECT (qtdemux, "found XdxT node"); /* collect the headers and store them in a stream list so that we can * send them out first */ qtdemux_parse_theora_extension (qtdemux, stream, xdxt); break; } default: break; } } GST_INFO_OBJECT (qtdemux, "type %" GST_FOURCC_FORMAT " caps %" GST_PTR_FORMAT, GST_FOURCC_ARGS (fourcc), stream->caps); } else if (stream->subtype == FOURCC_soun) { int version, samplesize; guint16 compression_id; offset = 32; if (len < 36) goto corrupt_file; version = QT_UINT32 (stsd_data + offset); stream->n_channels = QT_UINT16 (stsd_data + offset + 8); samplesize = QT_UINT16 (stsd_data + offset + 10); compression_id = QT_UINT16 (stsd_data + offset + 12); stream->rate = QT_FP32 (stsd_data + offset + 16); GST_LOG_OBJECT (qtdemux, "version/rev: %08x", version); GST_LOG_OBJECT (qtdemux, "vendor: %08x", QT_UINT32 (stsd_data + offset + 4)); GST_LOG_OBJECT (qtdemux, "n_channels: %d", stream->n_channels); GST_LOG_OBJECT (qtdemux, "sample_size: %d", samplesize); GST_LOG_OBJECT (qtdemux, "compression_id: %d", compression_id); GST_LOG_OBJECT (qtdemux, "packet size: %d", QT_UINT16 (stsd_data + offset + 14)); GST_LOG_OBJECT (qtdemux, "sample rate: %g", stream->rate); if (compression_id == 0xfffe) stream->sampled = TRUE; /* first assume uncompressed audio */ stream->bytes_per_sample = samplesize / 8; stream->samples_per_frame = stream->n_channels; stream->bytes_per_frame = stream->n_channels * stream->bytes_per_sample; stream->samples_per_packet = stream->samples_per_frame; stream->bytes_per_packet = stream->bytes_per_sample; offset = 52; switch (fourcc) { /* Yes, these have to be hard-coded */ case FOURCC_MAC6: { stream->samples_per_packet = 6; stream->bytes_per_packet = 1; stream->bytes_per_frame = 1 * stream->n_channels; stream->bytes_per_sample = 1; stream->samples_per_frame = 6 * stream->n_channels; break; } case FOURCC_MAC3: { stream->samples_per_packet = 3; stream->bytes_per_packet = 1; stream->bytes_per_frame = 1 * stream->n_channels; stream->bytes_per_sample = 1; stream->samples_per_frame = 3 * stream->n_channels; break; } case FOURCC_ima4: { stream->samples_per_packet = 64; stream->bytes_per_packet = 34; stream->bytes_per_frame = 34 * stream->n_channels; stream->bytes_per_sample = 2; stream->samples_per_frame = 64 * stream->n_channels; break; } case FOURCC_ulaw: case FOURCC_alaw: { stream->samples_per_packet = 1; stream->bytes_per_packet = 1; stream->bytes_per_frame = 1 * stream->n_channels; stream->bytes_per_sample = 1; stream->samples_per_frame = 1 * stream->n_channels; break; } case FOURCC_agsm: { stream->samples_per_packet = 160; stream->bytes_per_packet = 33; stream->bytes_per_frame = 33 * stream->n_channels; stream->bytes_per_sample = 2; stream->samples_per_frame = 160 * stream->n_channels; break; } default: break; } if (version == 0x00010000) { switch (fourcc) { case FOURCC_twos: case FOURCC_sowt: case FOURCC_raw_: break; default: { /* only parse extra decoding config for non-pcm audio */ stream->samples_per_packet = QT_UINT32 (stsd_data + offset); stream->bytes_per_packet = QT_UINT32 (stsd_data + offset + 4); stream->bytes_per_frame = QT_UINT32 (stsd_data + offset + 8); stream->bytes_per_sample = QT_UINT32 (stsd_data + offset + 12); GST_LOG_OBJECT (qtdemux, "samples/packet: %d", stream->samples_per_packet); GST_LOG_OBJECT (qtdemux, "bytes/packet: %d", stream->bytes_per_packet); GST_LOG_OBJECT (qtdemux, "bytes/frame: %d", stream->bytes_per_frame); GST_LOG_OBJECT (qtdemux, "bytes/sample: %d", stream->bytes_per_sample); if (!stream->sampled && stream->bytes_per_packet) { stream->samples_per_frame = (stream->bytes_per_frame / stream->bytes_per_packet) * stream->samples_per_packet; GST_LOG_OBJECT (qtdemux, "samples/frame: %d", stream->samples_per_frame); } break; } } } else if (version == 0x00020000) { union { gdouble fp; guint64 val; } qtfp; stream->samples_per_packet = QT_UINT32 (stsd_data + offset); qtfp.val = QT_UINT64 (stsd_data + offset + 4); stream->rate = qtfp.fp; stream->n_channels = QT_UINT32 (stsd_data + offset + 12); GST_LOG_OBJECT (qtdemux, "samples/packet: %d", stream->samples_per_packet); GST_LOG_OBJECT (qtdemux, "sample rate: %g", stream->rate); GST_LOG_OBJECT (qtdemux, "n_channels: %d", stream->n_channels); } else { GST_WARNING_OBJECT (qtdemux, "unknown version %08x", version); } stream->caps = qtdemux_audio_caps (qtdemux, stream, fourcc, NULL, 0, &codec); switch (fourcc) { case FOURCC_in24: { GNode *enda; GNode *in24; in24 = qtdemux_tree_get_child_by_type (stsd, FOURCC_in24); enda = qtdemux_tree_get_child_by_type (in24, FOURCC_enda); if (!enda) { wave = qtdemux_tree_get_child_by_type (in24, FOURCC_wave); if (wave) enda = qtdemux_tree_get_child_by_type (wave, FOURCC_enda); } if (enda) { gst_caps_set_simple (stream->caps, "endianness", G_TYPE_INT, G_LITTLE_ENDIAN, NULL); } break; } default: break; } if (codec) { list = gst_tag_list_new (); gst_tag_list_add (list, GST_TAG_MERGE_REPLACE, GST_TAG_AUDIO_CODEC, codec, NULL); g_free (codec); codec = NULL; } mp4a = qtdemux_tree_get_child_by_type (stsd, FOURCC_mp4a); wave = NULL; esds = NULL; if (mp4a) { wave = qtdemux_tree_get_child_by_type (mp4a, FOURCC_wave); if (wave) esds = qtdemux_tree_get_child_by_type (wave, FOURCC_esds); if (!esds) esds = qtdemux_tree_get_child_by_type (mp4a, FOURCC_esds); } if (esds) { gst_qtdemux_handle_esds (qtdemux, stream, esds, list); } else { switch (fourcc) { #if 0 /* FIXME: what is in the chunk? */ case FOURCC_QDMC: { gint len = QT_UINT32 (stsd_data); /* seems to be always = 116 = 0x74 */ break; } #endif case FOURCC_QDM2: { gint len = QT_UINT32 (stsd_data); if (len > 0x4C) { GstBuffer *buf = gst_buffer_new_and_alloc (len - 0x4C); memcpy (GST_BUFFER_DATA (buf), stsd_data + 0x4C, len - 0x4C); gst_caps_set_simple (stream->caps, "codec_data", GST_TYPE_BUFFER, buf, NULL); gst_buffer_unref (buf); } gst_caps_set_simple (stream->caps, "samplesize", G_TYPE_INT, samplesize, NULL); break; } case FOURCC_alac: { gint len = QT_UINT32 (stsd_data); if (len > 0x34) { GstBuffer *buf = gst_buffer_new_and_alloc (len - 0x34); memcpy (GST_BUFFER_DATA (buf), stsd_data + 0x34, len - 0x34); gst_caps_set_simple (stream->caps, "codec_data", GST_TYPE_BUFFER, buf, NULL); gst_buffer_unref (buf); } gst_caps_set_simple (stream->caps, "samplesize", G_TYPE_INT, samplesize, NULL); break; } case FOURCC_samr: { gint len = QT_UINT32 (stsd_data); if (len > 0x34) { GstBuffer *buf = gst_buffer_new_and_alloc (len - 0x34); memcpy (GST_BUFFER_DATA (buf), stsd_data + 0x34, len - 0x34); gst_caps_set_simple (stream->caps, "codec_data", GST_TYPE_BUFFER, buf, NULL); gst_buffer_unref (buf); } break; } default: break; } } GST_INFO_OBJECT (qtdemux, "type %" GST_FOURCC_FORMAT " caps %" GST_PTR_FORMAT, GST_FOURCC_ARGS (fourcc), stream->caps); } else if (stream->subtype == FOURCC_strm) { if (fourcc != FOURCC_rtsp) { GST_INFO_OBJECT (qtdemux, "unhandled stream type %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (fourcc)); goto unknown_stream; } stream->sampled = TRUE; } else if (stream->subtype == FOURCC_subp || stream->subtype == FOURCC_text) { stream->sampled = TRUE; offset = 16; stream->caps = qtdemux_sub_caps (qtdemux, stream, fourcc, stsd_data, &codec); if (codec) { list = gst_tag_list_new (); gst_tag_list_add (list, GST_TAG_MERGE_REPLACE, GST_TAG_SUBTITLE_CODEC, codec, NULL); g_free (codec); codec = NULL; } /* hunt for sort-of codec data */ switch (fourcc) { case FOURCC_mp4s: { guint len; const guint8 *data; /* look for palette */ /* target mp4s atom */ len = QT_UINT32 (stsd_data + offset); data = stsd_data + offset; /* verify sufficient length, * and esds present with decConfigDescr of expected size and position */ if ((len >= 106 + 8) && (QT_FOURCC (data + 8 + 8 + 4) == FOURCC_esds) && (QT_UINT16 (data + 8 + 40) == 0x0540)) { GstStructure *s; guint32 clut[16]; gint i; /* move to decConfigDescr data */ data = data + 8 + 42; for (i = 0; i < 16; i++) { clut[i] = QT_UINT32 (data); data += 4; } s = gst_structure_new ("application/x-gst-dvd", "event", G_TYPE_STRING, "dvd-spu-clut-change", "clut00", G_TYPE_INT, clut[0], "clut01", G_TYPE_INT, clut[1], "clut02", G_TYPE_INT, clut[2], "clut03", G_TYPE_INT, clut[3], "clut04", G_TYPE_INT, clut[4], "clut05", G_TYPE_INT, clut[5], "clut06", G_TYPE_INT, clut[6], "clut07", G_TYPE_INT, clut[7], "clut08", G_TYPE_INT, clut[8], "clut09", G_TYPE_INT, clut[9], "clut10", G_TYPE_INT, clut[10], "clut11", G_TYPE_INT, clut[11], "clut12", G_TYPE_INT, clut[12], "clut13", G_TYPE_INT, clut[13], "clut14", G_TYPE_INT, clut[14], "clut15", G_TYPE_INT, clut[15], NULL); /* store event and trigger custom processing */ stream->pending_event = gst_event_new_custom (GST_EVENT_CUSTOM_DOWNSTREAM, s); stream->need_process = TRUE; } break; } default: break; } } else { goto unknown_stream; } /* promote to sampled format */ if (stream->fourcc == FOURCC_samr) { /* force mono 8000 Hz for AMR */ stream->sampled = TRUE; stream->n_channels = 1; stream->rate = 8000; } else if (stream->fourcc == FOURCC_sawb) { /* force mono 16000 Hz for AMR-WB */ stream->sampled = TRUE; stream->n_channels = 1; stream->rate = 16000; } else if (stream->fourcc == FOURCC_mp4a) { stream->sampled = TRUE; } /* collect sample information */ if (!qtdemux_parse_samples (qtdemux, stream, stbl)) goto samples_failed; /* configure segments */ if (!qtdemux_parse_segments (qtdemux, stream, trak)) goto segments_failed; /* now we are ready to add the stream */ gst_qtdemux_add_stream (qtdemux, stream, list); return TRUE; /* ERRORS */ corrupt_file: { GST_ELEMENT_ERROR (qtdemux, STREAM, DECODE, (_("This file is corrupt and cannot be played.")), (NULL)); g_free (stream); return FALSE; } error_encrypted: { GST_ELEMENT_ERROR (qtdemux, STREAM, DECRYPT, (NULL), (NULL)); g_free (stream); return FALSE; } samples_failed: { /* we posted an error already */ g_free (stream); return FALSE; } segments_failed: { /* we posted an error already */ g_free (stream); return FALSE; } unknown_stream: { GST_INFO_OBJECT (qtdemux, "unknown subtype %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (stream->subtype)); g_free (stream); return TRUE; } } /* check if major or compatible brand is 3GP */ static inline gboolean qtdemux_is_brand_3gp (GstQTDemux * qtdemux, gboolean major) { if (major) { return ((qtdemux->major_brand & GST_MAKE_FOURCC (255, 255, 0, 0)) == GST_MAKE_FOURCC ('3', 'g', 0, 0)); } else if (qtdemux->comp_brands != NULL) { guint8 *data = GST_BUFFER_DATA (qtdemux->comp_brands); guint size = GST_BUFFER_SIZE (qtdemux->comp_brands); gboolean res = FALSE; while (size >= 4) { res = res || ((QT_FOURCC (data) & GST_MAKE_FOURCC (255, 255, 0, 0)) == GST_MAKE_FOURCC ('3', 'g', 0, 0)); data += 4; size -= 4; } return res; } else { return FALSE; } } /* check if tag is a spec'ed 3GP tag keyword storing a string */ static inline gboolean qtdemux_is_string_tag_3gp (GstQTDemux * qtdemux, guint32 fourcc) { return fourcc == FOURCC_cprt || fourcc == FOURCC_gnre || fourcc == FOURCC_titl || fourcc == FOURCC_dscp || fourcc == FOURCC_perf || fourcc == FOURCC_auth || fourcc == FOURCC_albm; } static void qtdemux_tag_add_location (GstQTDemux * qtdemux, const char *tag, const char *dummy, GNode * node) { const gchar *env_vars[] = { "GST_QT_TAG_ENCODING", "GST_TAG_ENCODING", NULL }; int offset; char *name; gchar *data; gdouble longitude, latitude, altitude; data = node->data; offset = 14; /* TODO: language code skipped */ name = gst_tag_freeform_string_to_utf8 (data + offset, -1, env_vars); if (!name) { /* do not alarm in trivial case, but bail out otherwise */ if (*(data + offset) != 0) { GST_DEBUG_OBJECT (qtdemux, "failed to convert %s tag to UTF-8, " "giving up", tag); } } else { gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, GST_TAG_GEO_LOCATION_NAME, name, NULL); offset += strlen (name); g_free (name); } /* +1 +1 = skip null-terminator and location role byte */ offset += 1 + 1; longitude = QT_FP32 (data + offset); offset += 4; latitude = QT_FP32 (data + offset); offset += 4; altitude = QT_FP32 (data + offset); /* one invalid means all are invalid */ if (longitude >= -180.0 && longitude <= 180.0 && latitude >= -90.0 && latitude <= 90.0) { gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, GST_TAG_GEO_LOCATION_LATITUDE, latitude, GST_TAG_GEO_LOCATION_LONGITUDE, longitude, GST_TAG_GEO_LOCATION_ELEVATION, altitude, NULL); } /* TODO: no GST_TAG_, so astronomical body and additional notes skipped */ } static void qtdemux_tag_add_year (GstQTDemux * qtdemux, const char *tag, const char *dummy, GNode * node) { guint16 y; GDate *date; y = QT_UINT16 ((guint8 *) node->data + 12); GST_DEBUG_OBJECT (qtdemux, "year: %u", y); date = g_date_new_dmy (1, 1, y); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag, date, NULL); g_date_free (date); } static void qtdemux_tag_add_classification (GstQTDemux * qtdemux, const char *tag, const char *dummy, GNode * node) { int offset; char *tag_str = NULL; guint8 *entity; guint16 table; offset = 12; entity = (guint8 *) node->data + offset; offset += 4; table = QT_UINT16 ((guint8 *) node->data + offset); /* Language code skipped */ offset += 4; /* Tag format: "XXXX://Y[YYYY]/classification info string" * XXXX: classification entity, fixed length 4 chars. * Y[YYYY]: classification table, max 5 chars. */ tag_str = g_strdup_printf ("----://%u/%s", table, (char *) node->data + offset); /* memcpy To be sure we're preserving byte order */ memcpy (tag_str, entity, 4); GST_DEBUG_OBJECT (qtdemux, "classification info: %s", tag_str); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_APPEND, tag, tag_str, NULL); g_free (tag_str); } static gboolean qtdemux_tag_add_str_full (GstQTDemux * qtdemux, const char *tag, const char *dummy, GNode * node) { const gchar *env_vars[] = { "GST_QT_TAG_ENCODING", "GST_TAG_ENCODING", NULL }; GNode *data; char *s; int len; guint32 type; int offset; gboolean ret = TRUE; data = qtdemux_tree_get_child_by_type (node, FOURCC_data); if (data) { len = QT_UINT32 (data->data); type = QT_UINT32 ((guint8 *) data->data + 8); if (type == 0x00000001) { s = gst_tag_freeform_string_to_utf8 ((char *) data->data + 16, len - 16, env_vars); if (s) { GST_DEBUG_OBJECT (qtdemux, "adding tag %s", GST_STR_NULL (s)); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag, s, NULL); g_free (s); } else { GST_DEBUG_OBJECT (qtdemux, "failed to convert %s tag to UTF-8", tag); } } } else { len = QT_UINT32 (node->data); type = QT_UINT32 ((guint8 *) node->data + 4); if ((type >> 24) == 0xa9) { /* Type starts with the (C) symbol, so the next 32 bits are * the language code, which we ignore */ offset = 12; GST_DEBUG_OBJECT (qtdemux, "found international text tag"); } else if (len > 14 && qtdemux_is_string_tag_3gp (qtdemux, QT_FOURCC ((guint8 *) node->data + 4))) { guint32 type = QT_UINT32 ((guint8 *) node->data + 8); /* we go for 3GP style encoding if major brands claims so, * or if no hope for data be ok UTF-8, and compatible 3GP brand present */ if (qtdemux_is_brand_3gp (qtdemux, TRUE) || (qtdemux_is_brand_3gp (qtdemux, FALSE) && ((type & 0x00FFFFFF) == 0x0) && (type >> 24 <= 0xF))) { offset = 14; /* 16-bit Language code is ignored here as well */ GST_DEBUG_OBJECT (qtdemux, "found 3gpp text tag"); } else { goto normal; } } else { normal: offset = 8; GST_DEBUG_OBJECT (qtdemux, "found normal text tag"); ret = FALSE; /* may have to fallback */ } s = gst_tag_freeform_string_to_utf8 ((char *) node->data + offset, len - offset, env_vars); if (s) { GST_DEBUG_OBJECT (qtdemux, "adding tag %s", GST_STR_NULL (s)); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag, s, NULL); g_free (s); ret = TRUE; } else { GST_DEBUG_OBJECT (qtdemux, "failed to convert %s tag to UTF-8", tag); } } return ret; } static void qtdemux_tag_add_str (GstQTDemux * qtdemux, const char *tag, const char *dummy, GNode * node) { qtdemux_tag_add_str_full (qtdemux, tag, dummy, node); } static void qtdemux_tag_add_keywords (GstQTDemux * qtdemux, const char *tag, const char *dummy, GNode * node) { const gchar *env_vars[] = { "GST_QT_TAG_ENCODING", "GST_TAG_ENCODING", NULL }; guint8 *data; char *s, *t, *k = NULL; int len; int offset; int count; /* first try normal string tag if major brand not 3GP */ if (!qtdemux_is_brand_3gp (qtdemux, TRUE)) { if (!qtdemux_tag_add_str_full (qtdemux, tag, dummy, node)) { /* hm, that did not work, maybe 3gpp storage in non-3gpp major brand; * let's try it 3gpp way after minor safety check */ data = node->data; if (QT_UINT32 (data) < 15 || !qtdemux_is_brand_3gp (qtdemux, FALSE)) return; } else return; } GST_DEBUG_OBJECT (qtdemux, "found 3gpp keyword tag"); data = node->data; len = QT_UINT32 (data); if (len < 15) goto short_read; count = QT_UINT8 (data + 14); offset = 15; for (; count; count--) { gint slen; if (offset + 1 > len) goto short_read; slen = QT_UINT8 (data + offset); offset += 1; if (offset + slen > len) goto short_read; s = gst_tag_freeform_string_to_utf8 ((char *) node->data + offset, slen, env_vars); if (s) { GST_DEBUG_OBJECT (qtdemux, "adding keyword %s", GST_STR_NULL (s)); if (k) { t = g_strjoin (",", k, s, NULL); g_free (s); g_free (k); k = t; } else { k = s; } } else { GST_DEBUG_OBJECT (qtdemux, "failed to convert keyword to UTF-8"); } offset += slen; } done: if (k) { GST_DEBUG_OBJECT (qtdemux, "adding tag %s", GST_STR_NULL (k)); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag, k, NULL); } g_free (k); return; /* ERRORS */ short_read: { GST_DEBUG_OBJECT (qtdemux, "short read parsing 3GP keywords"); goto done; } } static void qtdemux_tag_add_num (GstQTDemux * qtdemux, const char *tag1, const char *tag2, GNode * node) { GNode *data; int len; int type; int n1, n2; data = qtdemux_tree_get_child_by_type (node, FOURCC_data); if (data) { len = QT_UINT32 (data->data); type = QT_UINT32 ((guint8 *) data->data + 8); if (type == 0x00000000 && len >= 22) { n1 = QT_UINT16 ((guint8 *) data->data + 18); n2 = QT_UINT16 ((guint8 *) data->data + 20); if (n1 > 0) { GST_DEBUG_OBJECT (qtdemux, "adding tag %s=%d", tag1, n1); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag1, n1, NULL); } if (n2 > 0) { GST_DEBUG_OBJECT (qtdemux, "adding tag %s=%d", tag2, n2); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag2, n2, NULL); } } } } static void qtdemux_tag_add_tmpo (GstQTDemux * qtdemux, const char *tag1, const char *dummy, GNode * node) { GNode *data; int len; int type; int n1; data = qtdemux_tree_get_child_by_type (node, FOURCC_data); if (data) { len = QT_UINT32 (data->data); type = QT_UINT32 ((guint8 *) data->data + 8); GST_DEBUG_OBJECT (qtdemux, "have tempo tag, type=%d,len=%d", type, len); /* some files wrongly have a type 0x0f=15, but it should be 0x15 */ if ((type == 0x00000015 || type == 0x0000000f) && len >= 18) { n1 = QT_UINT16 ((guint8 *) data->data + 16); if (n1) { /* do not add bpm=0 */ GST_DEBUG_OBJECT (qtdemux, "adding tag %d", n1); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag1, (gdouble) n1, NULL); } } } } static void qtdemux_tag_add_covr (GstQTDemux * qtdemux, const char *tag1, const char *dummy, GNode * node) { GNode *data; int len; int type; GstBuffer *buf; data = qtdemux_tree_get_child_by_type (node, FOURCC_data); if (data) { len = QT_UINT32 (data->data); type = QT_UINT32 ((guint8 *) data->data + 8); GST_DEBUG_OBJECT (qtdemux, "have covr tag, type=%d,len=%d", type, len); if ((type == 0x0000000d || type == 0x0000000e) && len > 16) { if ((buf = gst_tag_image_data_to_image_buffer ((guint8 *) data->data + 16, len - 16, GST_TAG_IMAGE_TYPE_NONE))) { GST_DEBUG_OBJECT (qtdemux, "adding tag size %d", len - 16); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag1, buf, NULL); gst_buffer_unref (buf); } } } } static void qtdemux_tag_add_date (GstQTDemux * qtdemux, const char *tag, const char *dummy, GNode * node) { GNode *data; char *s; int len; int type; data = qtdemux_tree_get_child_by_type (node, FOURCC_data); if (data) { len = QT_UINT32 (data->data); type = QT_UINT32 ((guint8 *) data->data + 8); if (type == 0x00000001) { guint y, m = 1, d = 1; gint ret; s = g_strndup ((char *) data->data + 16, len - 16); GST_DEBUG_OBJECT (qtdemux, "adding date '%s'", s); ret = sscanf (s, "%u-%u-%u", &y, &m, &d); if (ret >= 1 && y > 1500 && y < 3000) { GDate *date; date = g_date_new_dmy (d, m, y); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag, date, NULL); g_date_free (date); } else { GST_DEBUG_OBJECT (qtdemux, "could not parse date string '%s'", s); } g_free (s); } } } static void qtdemux_tag_add_gnre (GstQTDemux * qtdemux, const char *tag, const char *dummy, GNode * node) { GNode *data; data = qtdemux_tree_get_child_by_type (node, FOURCC_data); /* re-route to normal string tag if major brand says so * or no data atom and compatible brand suggests so */ if (qtdemux_is_brand_3gp (qtdemux, TRUE) || (qtdemux_is_brand_3gp (qtdemux, FALSE) && !data)) { qtdemux_tag_add_str (qtdemux, tag, dummy, node); return; } if (data) { guint len, type, n; len = QT_UINT32 (data->data); type = QT_UINT32 ((guint8 *) data->data + 8); if (type == 0x00000000 && len >= 18) { n = QT_UINT16 ((guint8 *) data->data + 16); if (n > 0) { const gchar *genre; genre = gst_tag_id3_genre_get (n - 1); if (genre != NULL) { GST_DEBUG_OBJECT (qtdemux, "adding %d [%s]", n, genre); gst_tag_list_add (qtdemux->tag_list, GST_TAG_MERGE_REPLACE, tag, genre, NULL); } } } } } typedef void (*GstQTDemuxAddTagFunc) (GstQTDemux * demux, const char *tag, const char *tag_bis, GNode * node); static const struct { guint32 fourcc; const gchar *gst_tag; const gchar *gst_tag_bis; const GstQTDemuxAddTagFunc func; } add_funcs[] = { { FOURCC__nam, GST_TAG_TITLE, NULL, qtdemux_tag_add_str}, { FOURCC_titl, GST_TAG_TITLE, NULL, qtdemux_tag_add_str}, { FOURCC__grp, GST_TAG_ARTIST, NULL, qtdemux_tag_add_str}, { FOURCC__wrt, GST_TAG_COMPOSER, NULL, qtdemux_tag_add_str}, { FOURCC__ART, GST_TAG_ARTIST, NULL, qtdemux_tag_add_str}, { FOURCC_perf, GST_TAG_ARTIST, NULL, qtdemux_tag_add_str}, { FOURCC_auth, GST_TAG_COMPOSER, NULL, qtdemux_tag_add_str}, { FOURCC__alb, GST_TAG_ALBUM, NULL, qtdemux_tag_add_str}, { FOURCC_albm, GST_TAG_ALBUM, NULL, qtdemux_tag_add_str}, { FOURCC_cprt, GST_TAG_COPYRIGHT, NULL, qtdemux_tag_add_str}, { FOURCC__cpy, GST_TAG_COPYRIGHT, NULL, qtdemux_tag_add_str}, { FOURCC__cmt, GST_TAG_COMMENT, NULL, qtdemux_tag_add_str}, { FOURCC__des, GST_TAG_DESCRIPTION, NULL, qtdemux_tag_add_str}, { FOURCC_dscp, GST_TAG_DESCRIPTION, NULL, qtdemux_tag_add_str}, { FOURCC__day, GST_TAG_DATE, NULL, qtdemux_tag_add_date}, { FOURCC_yrrc, GST_TAG_DATE, NULL, qtdemux_tag_add_year}, { FOURCC__too, GST_TAG_COMMENT, NULL, qtdemux_tag_add_str}, { FOURCC__inf, GST_TAG_COMMENT, NULL, qtdemux_tag_add_str}, { FOURCC_trkn, GST_TAG_TRACK_NUMBER, GST_TAG_TRACK_COUNT, qtdemux_tag_add_num}, { FOURCC_disk, GST_TAG_ALBUM_VOLUME_NUMBER, GST_TAG_ALBUM_VOLUME_COUNT, qtdemux_tag_add_num}, { FOURCC_disc, GST_TAG_ALBUM_VOLUME_NUMBER, GST_TAG_ALBUM_VOLUME_COUNT, qtdemux_tag_add_num}, { FOURCC__gen, GST_TAG_GENRE, NULL, qtdemux_tag_add_str}, { FOURCC_gnre, GST_TAG_GENRE, NULL, qtdemux_tag_add_gnre}, { FOURCC_tmpo, GST_TAG_BEATS_PER_MINUTE, NULL, qtdemux_tag_add_tmpo}, { FOURCC_covr, GST_TAG_PREVIEW_IMAGE, NULL, qtdemux_tag_add_covr}, { FOURCC_kywd, GST_TAG_KEYWORDS, NULL, qtdemux_tag_add_keywords}, { FOURCC_keyw, GST_TAG_KEYWORDS, NULL, qtdemux_tag_add_str}, { FOURCC__enc, GST_TAG_ENCODER, NULL, qtdemux_tag_add_str}, { FOURCC_loci, GST_TAG_GEO_LOCATION_NAME, NULL, qtdemux_tag_add_location}, { FOURCC_clsf, GST_QT_DEMUX_CLASSIFICATION_TAG, NULL, qtdemux_tag_add_classification} }; static void qtdemux_tag_add_blob (GNode * node, GstQTDemux * demux) { gint len; guint8 *data; GstBuffer *buf; gchar *media_type, *style; GstCaps *caps; guint i; guint8 ndata[4]; data = node->data; len = QT_UINT32 (data); buf = gst_buffer_new_and_alloc (len); memcpy (GST_BUFFER_DATA (buf), data, len); /* heuristic to determine style of tag */ if (QT_FOURCC (data + 4) == FOURCC_____ || (len > 8 + 12 && QT_FOURCC (data + 12) == FOURCC_data)) style = "itunes"; else if (demux->major_brand == FOURCC_qt__) style = "quicktime"; /* fall back to assuming iso/3gp tag style */ else style = "iso"; /* santize the name for the caps. */ for (i = 0; i < 4; i++) { guint8 d = data[4 + i]; if (g_ascii_isalnum (d)) ndata[i] = g_ascii_tolower (d); else ndata[i] = '_'; } media_type = g_strdup_printf ("application/x-gst-qt-%c%c%c%c-tag", ndata[0], ndata[1], ndata[2], ndata[3]); GST_DEBUG_OBJECT (demux, "media type %s", media_type); caps = gst_caps_new_simple (media_type, "style", G_TYPE_STRING, style, NULL); gst_buffer_set_caps (buf, caps); gst_caps_unref (caps); g_free (media_type); GST_DEBUG_OBJECT (demux, "adding private tag; size %d, caps %" GST_PTR_FORMAT, GST_BUFFER_SIZE (buf), caps); gst_tag_list_add (demux->tag_list, GST_TAG_MERGE_APPEND, GST_QT_DEMUX_PRIVATE_TAG, buf, NULL); gst_buffer_unref (buf); } static void qtdemux_parse_udta (GstQTDemux * qtdemux, GNode * udta) { GNode *meta; GNode *ilst; GNode *node; gint i; meta = qtdemux_tree_get_child_by_type (udta, FOURCC_meta); if (meta != NULL) { ilst = qtdemux_tree_get_child_by_type (meta, FOURCC_ilst); if (ilst == NULL) { GST_LOG_OBJECT (qtdemux, "no ilst"); return; } } else { ilst = udta; GST_LOG_OBJECT (qtdemux, "no meta so using udta itself"); } GST_DEBUG_OBJECT (qtdemux, "new tag list"); qtdemux->tag_list = gst_tag_list_new (); for (i = 0; i < G_N_ELEMENTS (add_funcs); ++i) { node = qtdemux_tree_get_child_by_type (ilst, add_funcs[i].fourcc); if (node) { add_funcs[i].func (qtdemux, add_funcs[i].gst_tag, add_funcs[i].gst_tag_bis, node); g_node_destroy (node); } } /* parsed nodes have been removed, pass along remainder as blob */ g_node_children_foreach (ilst, G_TRAVERSE_ALL, (GNodeForeachFunc) qtdemux_tag_add_blob, qtdemux); } typedef struct { GstStructure *structure; /* helper for sort function */ gchar *location; guint min_req_bitrate; guint min_req_qt_version; } GstQtReference; static gint qtdemux_redirects_sort_func (gconstpointer a, gconstpointer b) { GstQtReference *ref_a = (GstQtReference *) a; GstQtReference *ref_b = (GstQtReference *) b; if (ref_b->min_req_qt_version != ref_a->min_req_qt_version) return ref_b->min_req_qt_version - ref_a->min_req_qt_version; /* known bitrates go before unknown; higher bitrates go first */ return ref_b->min_req_bitrate - ref_a->min_req_bitrate; } /* sort the redirects and post a message for the application. */ static void qtdemux_process_redirects (GstQTDemux * qtdemux, GList * references) { GstQtReference *best; GstStructure *s; GstMessage *msg; GValue list_val = { 0, }; GList *l; g_assert (references != NULL); references = g_list_sort (references, qtdemux_redirects_sort_func); best = (GstQtReference *) references->data; g_value_init (&list_val, GST_TYPE_LIST); for (l = references; l != NULL; l = l->next) { GstQtReference *ref = (GstQtReference *) l->data; GValue struct_val = { 0, }; ref->structure = gst_structure_new ("redirect", "new-location", G_TYPE_STRING, ref->location, NULL); if (ref->min_req_bitrate > 0) { gst_structure_set (ref->structure, "minimum-bitrate", G_TYPE_INT, ref->min_req_bitrate, NULL); } g_value_init (&struct_val, GST_TYPE_STRUCTURE); g_value_set_boxed (&struct_val, ref->structure); gst_value_list_append_value (&list_val, &struct_val); g_value_unset (&struct_val); /* don't free anything here yet, since we need best->structure below */ } g_assert (best != NULL); s = gst_structure_copy (best->structure); if (g_list_length (references) > 1) { gst_structure_set_value (s, "locations", &list_val); } g_value_unset (&list_val); for (l = references; l != NULL; l = l->next) { GstQtReference *ref = (GstQtReference *) l->data; gst_structure_free (ref->structure); g_free (ref->location); g_free (ref); } g_list_free (references); GST_INFO_OBJECT (qtdemux, "posting redirect message: %" GST_PTR_FORMAT, s); msg = gst_message_new_element (GST_OBJECT_CAST (qtdemux), s); gst_element_post_message (GST_ELEMENT_CAST (qtdemux), msg); } /* look for redirect nodes, collect all redirect information and * process it. */ static gboolean qtdemux_parse_redirects (GstQTDemux * qtdemux) { GNode *rmra, *rmda, *rdrf; rmra = qtdemux_tree_get_child_by_type (qtdemux->moov_node, FOURCC_rmra); if (rmra) { GList *redirects = NULL; rmda = qtdemux_tree_get_child_by_type (rmra, FOURCC_rmda); while (rmda) { GstQtReference ref = { NULL, NULL, 0, 0 }; GNode *rmdr, *rmvc; if ((rmdr = qtdemux_tree_get_child_by_type (rmda, FOURCC_rmdr))) { ref.min_req_bitrate = QT_UINT32 ((guint8 *) rmdr->data + 12); GST_LOG_OBJECT (qtdemux, "data rate atom, required bitrate = %u", ref.min_req_bitrate); } if ((rmvc = qtdemux_tree_get_child_by_type (rmda, FOURCC_rmvc))) { guint32 package = QT_FOURCC ((guint8 *) rmvc->data + 12); guint version = QT_UINT32 ((guint8 *) rmvc->data + 16); #ifndef GST_DISABLE_GST_DEBUG guint bitmask = QT_UINT32 ((guint8 *) rmvc->data + 20); #endif guint check_type = QT_UINT16 ((guint8 *) rmvc->data + 24); GST_LOG_OBJECT (qtdemux, "version check atom [%" GST_FOURCC_FORMAT "], version=0x%08x" ", mask=%08x, check_type=%u", GST_FOURCC_ARGS (package), version, bitmask, check_type); if (package == FOURCC_qtim && check_type == 0) { ref.min_req_qt_version = version; } } rdrf = qtdemux_tree_get_child_by_type (rmda, FOURCC_rdrf); if (rdrf) { guint32 ref_type; guint8 *ref_data; ref_type = QT_FOURCC ((guint8 *) rdrf->data + 12); ref_data = (guint8 *) rdrf->data + 20; if (ref_type == FOURCC_alis) { guint record_len, record_version, fn_len; /* MacOSX alias record, google for alias-layout.txt */ record_len = QT_UINT16 (ref_data + 4); record_version = QT_UINT16 (ref_data + 4 + 2); fn_len = QT_UINT8 (ref_data + 50); if (record_len > 50 && record_version == 2 && fn_len > 0) { ref.location = g_strndup ((gchar *) ref_data + 51, fn_len); } } else if (ref_type == FOURCC_url_) { ref.location = g_strdup ((gchar *) ref_data); } else { GST_DEBUG_OBJECT (qtdemux, "unknown rdrf reference type %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (ref_type)); } if (ref.location != NULL) { GST_INFO_OBJECT (qtdemux, "New location: %s", ref.location); redirects = g_list_prepend (redirects, g_memdup (&ref, sizeof (ref))); } else { GST_WARNING_OBJECT (qtdemux, "Failed to extract redirect location from rdrf atom"); } } /* look for others */ rmda = qtdemux_tree_get_sibling_by_type (rmda, FOURCC_rmda); } if (redirects != NULL) { qtdemux_process_redirects (qtdemux, redirects); } } return TRUE; } static GstTagList * qtdemux_add_container_format (GstQTDemux * qtdemux, GstTagList * tags) { const gchar *fmt; if (tags == NULL) tags = gst_tag_list_new (); if (qtdemux->major_brand == FOURCC_mjp2) fmt = "Motion JPEG 2000"; else if ((qtdemux->major_brand & 0xffff) == GST_MAKE_FOURCC ('3', 'g', 0, 0)) fmt = "3GP"; else if (qtdemux->major_brand == FOURCC_qt__) fmt = "Quicktime"; else fmt = "ISO MP4/M4A"; GST_LOG_OBJECT (qtdemux, "mapped %" GST_FOURCC_FORMAT " to '%s'", GST_FOURCC_ARGS (qtdemux->major_brand), fmt); gst_tag_list_add (tags, GST_TAG_MERGE_REPLACE, GST_TAG_CONTAINER_FORMAT, fmt, NULL); return tags; } /* we have read th complete moov node now. * This function parses all of the relevant info, creates the traks and * prepares all data structures for playback */ static gboolean qtdemux_parse_tree (GstQTDemux * qtdemux) { GNode *mvhd; GNode *trak; GNode *udta; gint64 duration; mvhd = qtdemux_tree_get_child_by_type (qtdemux->moov_node, FOURCC_mvhd); if (mvhd == NULL) { GST_LOG_OBJECT (qtdemux, "No mvhd node found, looking for redirects."); return qtdemux_parse_redirects (qtdemux); } qtdemux->timescale = QT_UINT32 ((guint8 *) mvhd->data + 20); qtdemux->duration = QT_UINT32 ((guint8 *) mvhd->data + 24); GST_INFO_OBJECT (qtdemux, "timescale: %u", qtdemux->timescale); GST_INFO_OBJECT (qtdemux, "duration: %u", qtdemux->duration); /* set duration in the segment info */ gst_qtdemux_get_duration (qtdemux, &duration); gst_segment_set_duration (&qtdemux->segment, GST_FORMAT_TIME, duration); /* parse all traks */ trak = qtdemux_tree_get_child_by_type (qtdemux->moov_node, FOURCC_trak); while (trak) { qtdemux_parse_trak (qtdemux, trak); /* iterate all siblings */ trak = qtdemux_tree_get_sibling_by_type (trak, FOURCC_trak); } gst_element_no_more_pads (GST_ELEMENT_CAST (qtdemux)); /* find and push tags, we do this after adding the pads so we can push the * tags downstream as well. */ udta = qtdemux_tree_get_child_by_type (qtdemux->moov_node, FOURCC_udta); if (udta) { qtdemux_parse_udta (qtdemux, udta); } else { GST_LOG_OBJECT (qtdemux, "No udta node found."); } qtdemux->tag_list = qtdemux_add_container_format (qtdemux, qtdemux->tag_list); GST_INFO_OBJECT (qtdemux, "posting global tags: %" GST_PTR_FORMAT, qtdemux->tag_list); /* post now, send event on pads later */ gst_element_post_message (GST_ELEMENT (qtdemux), gst_message_new_tag (GST_OBJECT (qtdemux), gst_tag_list_copy (qtdemux->tag_list))); return TRUE; } /* taken from ffmpeg */ static unsigned int get_size (guint8 * ptr, guint8 ** end) { int count = 4; int len = 0; while (count--) { int c = *ptr; ptr++; len = (len << 7) | (c & 0x7f); if (!(c & 0x80)) break; } if (end) *end = ptr; return len; } /* this can change the codec originally present in @list */ static void gst_qtdemux_handle_esds (GstQTDemux * qtdemux, QtDemuxStream * stream, GNode * esds, GstTagList * list) { int len = QT_UINT32 (esds->data); guint8 *ptr = esds->data; guint8 *end = ptr + len; int tag; guint8 *data_ptr = NULL; int data_len = 0; guint8 object_type_id = 0; char *codec_name = NULL; GstCaps *caps = NULL; GST_MEMDUMP_OBJECT (qtdemux, "esds", ptr, len); ptr += 8; GST_DEBUG_OBJECT (qtdemux, "version/flags = %08x", QT_UINT32 (ptr)); ptr += 4; while (ptr < end) { tag = QT_UINT8 (ptr); GST_DEBUG_OBJECT (qtdemux, "tag = %02x", tag); ptr++; len = get_size (ptr, &ptr); GST_DEBUG_OBJECT (qtdemux, "len = %d", len); switch (tag) { case 0x03: GST_DEBUG_OBJECT (qtdemux, "ID %04x", QT_UINT16 (ptr)); GST_DEBUG_OBJECT (qtdemux, "priority %04x", QT_UINT8 (ptr + 2)); ptr += 3; break; case 0x04: object_type_id = QT_UINT8 (ptr); GST_DEBUG_OBJECT (qtdemux, "object_type_id %02x", object_type_id); GST_DEBUG_OBJECT (qtdemux, "stream_type %02x", QT_UINT8 (ptr + 1)); GST_DEBUG_OBJECT (qtdemux, "buffer_size_db %02x", QT_UINT24 (ptr + 2)); GST_DEBUG_OBJECT (qtdemux, "max bitrate %d", QT_UINT32 (ptr + 5)); GST_DEBUG_OBJECT (qtdemux, "avg bitrate %d", QT_UINT32 (ptr + 9)); ptr += 13; break; case 0x05: GST_MEMDUMP_OBJECT (qtdemux, "data", ptr, len); data_ptr = ptr; data_len = len; ptr += len; break; case 0x06: GST_DEBUG_OBJECT (qtdemux, "data %02x", QT_UINT8 (ptr)); ptr += 1; break; default: GST_ERROR_OBJECT (qtdemux, "parse error"); break; } } /* object_type_id in the esds atom in mp4a and mp4v tells us which codec is * in use, and should also be used to override some other parameters for some * codecs. */ switch (object_type_id) { case 0x20: /* MPEG-4 */ break; /* Nothing special needed here */ case 0x21: /* H.264 */ codec_name = "H.264 / AVC"; caps = gst_caps_new_simple ("video/x-h264", NULL); break; case 0x40: /* AAC (any) */ case 0x66: /* AAC Main */ case 0x67: /* AAC LC */ case 0x68: /* AAC SSR */ /* Override channels and rate based on the codec_data, as it's often * wrong. */ if (data_ptr && data_len >= 2) { guint channels, rateindex; int rates[] = { 96000, 88200, 64000, 48000, 44100, 32000, 24000, 22050, 16000, 12000, 11025, 8000 }; channels = (data_ptr[1] & 0x7f) >> 3; if (channels <= 7) { stream->n_channels = channels; } rateindex = ((data_ptr[0] & 0x7) << 1) | ((data_ptr[1] & 0x80) >> 7); if (rateindex < sizeof (rates) / sizeof (*rates)) { stream->rate = rates[rateindex]; } } break; case 0x60: /* MPEG-2, various profiles */ case 0x61: case 0x62: case 0x63: case 0x64: case 0x65: codec_name = "MPEG-2 video"; gst_caps_unref (stream->caps); stream->caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, 2, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case 0x69: /* MP3 has two different values, accept either */ case 0x6B: /* change to mpeg1 layer 3 audio */ gst_caps_set_simple (stream->caps, "layer", G_TYPE_INT, 3, "mpegversion", G_TYPE_INT, 1, NULL); codec_name = "MPEG-1 layer 3"; break; case 0x6A: /* MPEG-1 */ codec_name = "MPEG-1 video"; gst_caps_unref (stream->caps); stream->caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, 1, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case 0x6C: /* MJPEG */ caps = gst_caps_new_simple ("image/jpeg", NULL); codec_name = "Motion-JPEG"; break; case 0x6D: /* PNG */ caps = gst_caps_new_simple ("image/png", NULL); codec_name = "PNG still images"; break; case 0x6E: /* JPEG2000 */ codec_name = "JPEG-2000"; caps = gst_caps_new_simple ("image/x-j2c", "fields", G_TYPE_INT, 1, NULL); break; case 0xA4: /* Dirac */ codec_name = "Dirac"; caps = gst_caps_new_simple ("video/x-dirac", NULL); break; case 0xA5: /* AC3 */ codec_name = "AC-3 audio"; caps = gst_caps_new_simple ("audio/x-ac3", NULL); break; case 0xE1: /* QCELP */ /* QCELP, the codec_data is a riff tag (little endian) with * more info (http://ftp.3gpp2.org/TSGC/Working/2003/2003-05-SanDiego/TSG-C-2003-05-San%20Diego/WG1/SWG12/C12-20030512-006%20=%20C12-20030217-015_Draft_Baseline%20Text%20of%20FFMS_R2.doc). */ caps = gst_caps_new_simple ("audio/qcelp", NULL); codec_name = "QCELP"; break; default: break; } /* If we have a replacement caps, then change our caps for this stream */ if (caps) { gst_caps_unref (stream->caps); stream->caps = caps; } if (codec_name && list) gst_tag_list_add (list, GST_TAG_MERGE_REPLACE, GST_TAG_AUDIO_CODEC, codec_name, NULL); /* Add the codec_data attribute to caps, if we have it */ if (data_ptr) { GstBuffer *buffer; buffer = gst_buffer_new_and_alloc (data_len); memcpy (GST_BUFFER_DATA (buffer), data_ptr, data_len); GST_DEBUG_OBJECT (qtdemux, "setting codec_data from esds"); GST_MEMDUMP_OBJECT (qtdemux, "codec_data from esds", data_ptr, data_len); gst_caps_set_simple (stream->caps, "codec_data", GST_TYPE_BUFFER, buffer, NULL); gst_buffer_unref (buffer); } } #define _codec(name) \ do { \ if (codec_name) { \ *codec_name = g_strdup (name); \ } \ } while (0) static GstCaps * qtdemux_video_caps (GstQTDemux * qtdemux, QtDemuxStream * stream, guint32 fourcc, const guint8 * stsd_data, gchar ** codec_name) { GstCaps *caps; const GstStructure *s; const gchar *name; switch (fourcc) { case GST_MAKE_FOURCC ('p', 'n', 'g', ' '): _codec ("PNG still images"); caps = gst_caps_new_simple ("image/png", NULL); break; case GST_MAKE_FOURCC ('j', 'p', 'e', 'g'): _codec ("JPEG still images"); caps = gst_caps_new_simple ("image/jpeg", NULL); break; case GST_MAKE_FOURCC ('m', 'j', 'p', 'a'): case GST_MAKE_FOURCC ('A', 'V', 'D', 'J'): case GST_MAKE_FOURCC ('M', 'J', 'P', 'G'): _codec ("Motion-JPEG"); caps = gst_caps_new_simple ("image/jpeg", NULL); break; case GST_MAKE_FOURCC ('m', 'j', 'p', 'b'): _codec ("Motion-JPEG format B"); caps = gst_caps_new_simple ("video/x-mjpeg-b", NULL); break; case GST_MAKE_FOURCC ('m', 'j', 'p', '2'): _codec ("JPEG-2000"); /* override to what it should be according to spec, avoid palette_data */ stream->bits_per_sample = 24; caps = gst_caps_new_simple ("image/x-j2c", "fields", G_TYPE_INT, 1, NULL); break; case GST_MAKE_FOURCC ('S', 'V', 'Q', '3'): _codec ("Sorensen video v.3"); caps = gst_caps_new_simple ("video/x-svq", "svqversion", G_TYPE_INT, 3, NULL); break; case GST_MAKE_FOURCC ('s', 'v', 'q', 'i'): case GST_MAKE_FOURCC ('S', 'V', 'Q', '1'): _codec ("Sorensen video v.1"); caps = gst_caps_new_simple ("video/x-svq", "svqversion", G_TYPE_INT, 1, NULL); break; case GST_MAKE_FOURCC ('r', 'a', 'w', ' '): { guint16 bps; _codec ("Raw RGB video"); bps = QT_UINT16 (stsd_data + 98); /* set common stuff */ caps = gst_caps_new_simple ("video/x-raw-rgb", "endianness", G_TYPE_INT, G_BYTE_ORDER, "depth", G_TYPE_INT, bps, NULL); switch (bps) { case 15: gst_caps_set_simple (caps, "bpp", G_TYPE_INT, 16, "endianness", G_TYPE_INT, G_BIG_ENDIAN, "red_mask", G_TYPE_INT, 0x7c00, "green_mask", G_TYPE_INT, 0x03e0, "blue_mask", G_TYPE_INT, 0x001f, NULL); break; case 16: gst_caps_set_simple (caps, "bpp", G_TYPE_INT, 16, "endianness", G_TYPE_INT, G_BIG_ENDIAN, "red_mask", G_TYPE_INT, 0xf800, "green_mask", G_TYPE_INT, 0x07e0, "blue_mask", G_TYPE_INT, 0x001f, NULL); break; case 24: gst_caps_set_simple (caps, "bpp", G_TYPE_INT, 24, "endianness", G_TYPE_INT, G_BIG_ENDIAN, "red_mask", G_TYPE_INT, 0xff0000, "green_mask", G_TYPE_INT, 0x00ff00, "blue_mask", G_TYPE_INT, 0x0000ff, NULL); break; case 32: gst_caps_set_simple (caps, "bpp", G_TYPE_INT, 32, "endianness", G_TYPE_INT, G_BIG_ENDIAN, "alpha_mask", G_TYPE_INT, 0xff000000, "red_mask", G_TYPE_INT, 0x00ff0000, "green_mask", G_TYPE_INT, 0x0000ff00, "blue_mask", G_TYPE_INT, 0x000000ff, NULL); break; default: /* unknown */ break; } break; } case GST_MAKE_FOURCC ('y', 'v', '1', '2'): _codec ("Raw planar YUV 4:2:0"); caps = gst_caps_new_simple ("video/x-raw-yuv", "format", GST_TYPE_FOURCC, GST_MAKE_FOURCC ('I', '4', '2', '0'), NULL); break; case GST_MAKE_FOURCC ('y', 'u', 'v', '2'): case GST_MAKE_FOURCC ('Y', 'u', 'v', '2'): _codec ("Raw packed YUV 4:2:2"); caps = gst_caps_new_simple ("video/x-raw-yuv", "format", GST_TYPE_FOURCC, GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'), NULL); break; case GST_MAKE_FOURCC ('2', 'v', 'u', 'y'): _codec ("Raw packed YUV 4:2:0"); caps = gst_caps_new_simple ("video/x-raw-yuv", "format", GST_TYPE_FOURCC, GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'), NULL); break; case GST_MAKE_FOURCC ('v', '2', '1', '0'): _codec ("Raw packed YUV 10-bit 4:2:2"); caps = gst_caps_new_simple ("video/x-raw-yuv", "format", GST_TYPE_FOURCC, GST_MAKE_FOURCC ('v', '2', '1', '0'), NULL); break; case GST_MAKE_FOURCC ('m', 'p', 'e', 'g'): case GST_MAKE_FOURCC ('m', 'p', 'g', '1'): _codec ("MPEG-1 video"); caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, 1, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MAKE_FOURCC ('h', 'd', 'v', '1'): // HDV 720p30 case GST_MAKE_FOURCC ('h', 'd', 'v', '2'): // HDV 1080i60 case GST_MAKE_FOURCC ('h', 'd', 'v', '3'): // HDV 1080i50 case GST_MAKE_FOURCC ('h', 'd', 'v', '5'): // HDV 720p25 case GST_MAKE_FOURCC ('h', 'd', 'v', '6'): // HDV 1080i60 case GST_MAKE_FOURCC ('m', 'x', '5', 'n'): // MPEG2 IMX NTSC 525/60 50mb/s produced by FCP case GST_MAKE_FOURCC ('m', 'x', '5', 'p'): // MPEG2 IMX PAL 625/60 50mb/s produced by FCP case GST_MAKE_FOURCC ('m', 'x', '4', 'n'): // MPEG2 IMX NTSC 525/60 40mb/s produced by FCP case GST_MAKE_FOURCC ('m', 'x', '4', 'p'): // MPEG2 IMX PAL 625/60 40mb/s produced by FCP case GST_MAKE_FOURCC ('m', 'x', '3', 'n'): // MPEG2 IMX NTSC 525/60 30mb/s produced by FCP case GST_MAKE_FOURCC ('m', 'x', '3', 'p'): // MPEG2 IMX PAL 625/50 30mb/s produced by FCP case GST_MAKE_FOURCC ('x', 'd', 'v', '2'): // XDCAM HD 1080i60 case GST_MAKE_FOURCC ('A', 'V', 'm', 'p'): // AVID IMX PAL case GST_MAKE_FOURCC ('m', 'p', 'g', '2'): // AVID IMX PAL _codec ("MPEG-2 video"); caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, 2, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MAKE_FOURCC ('g', 'i', 'f', ' '): _codec ("GIF still images"); caps = gst_caps_new_simple ("image/gif", NULL); break; case GST_MAKE_FOURCC ('h', '2', '6', '3'): case GST_MAKE_FOURCC ('H', '2', '6', '3'): case GST_MAKE_FOURCC ('s', '2', '6', '3'): case GST_MAKE_FOURCC ('U', '2', '6', '3'): _codec ("H.263"); /* ffmpeg uses the height/width props, don't know why */ caps = gst_caps_new_simple ("video/x-h263", NULL); break; case GST_MAKE_FOURCC ('m', 'p', '4', 'v'): _codec ("MPEG-4 video"); caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, 4, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MAKE_FOURCC ('3', 'i', 'v', 'd'): case GST_MAKE_FOURCC ('3', 'I', 'V', 'D'): _codec ("Microsoft MPEG-4 4.3"); /* FIXME? */ caps = gst_caps_new_simple ("video/x-msmpeg", "msmpegversion", G_TYPE_INT, 43, NULL); break; case GST_MAKE_FOURCC ('3', 'I', 'V', '1'): case GST_MAKE_FOURCC ('3', 'I', 'V', '2'): _codec ("3ivX video"); caps = gst_caps_new_simple ("video/x-3ivx", NULL); break; case GST_MAKE_FOURCC ('D', 'I', 'V', '3'): _codec ("DivX 3"); caps = gst_caps_new_simple ("video/x-divx", "divxversion", G_TYPE_INT, 3, NULL); break; case GST_MAKE_FOURCC ('D', 'I', 'V', 'X'): case GST_MAKE_FOURCC ('d', 'i', 'v', 'x'): _codec ("DivX 4"); caps = gst_caps_new_simple ("video/x-divx", "divxversion", G_TYPE_INT, 4, NULL); break; case GST_MAKE_FOURCC ('D', 'X', '5', '0'): _codec ("DivX 5"); caps = gst_caps_new_simple ("video/x-divx", "divxversion", G_TYPE_INT, 5, NULL); break; case GST_MAKE_FOURCC ('X', 'V', 'I', 'D'): case GST_MAKE_FOURCC ('x', 'v', 'i', 'd'): _codec ("XVID MPEG-4"); caps = gst_caps_new_simple ("video/x-xvid", NULL); break; case GST_MAKE_FOURCC ('F', 'M', 'P', '4'): case GST_MAKE_FOURCC ('U', 'M', 'P', '4'): caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, 4, NULL); if (codec_name) *codec_name = g_strdup ("FFmpeg MPEG-4"); break; case GST_MAKE_FOURCC ('c', 'v', 'i', 'd'): _codec ("Cinepak"); caps = gst_caps_new_simple ("video/x-cinepak", NULL); break; case GST_MAKE_FOURCC ('q', 'd', 'r', 'w'): _codec ("Apple QuickDraw"); caps = gst_caps_new_simple ("video/x-qdrw", NULL); break; case GST_MAKE_FOURCC ('r', 'p', 'z', 'a'): _codec ("Apple video"); caps = gst_caps_new_simple ("video/x-apple-video", NULL); break; case GST_MAKE_FOURCC ('a', 'v', 'c', '1'): _codec ("H.264 / AVC"); caps = gst_caps_new_simple ("video/x-h264", NULL); break; case GST_MAKE_FOURCC ('r', 'l', 'e', ' '): _codec ("Run-length encoding"); caps = gst_caps_new_simple ("video/x-rle", "layout", G_TYPE_STRING, "quicktime", NULL); break; case GST_MAKE_FOURCC ('i', 'v', '3', '2'): _codec ("Indeo Video 3"); caps = gst_caps_new_simple ("video/x-indeo", "indeoversion", G_TYPE_INT, 3, NULL); break; case GST_MAKE_FOURCC ('I', 'V', '4', '1'): case GST_MAKE_FOURCC ('i', 'v', '4', '1'): _codec ("Intel Video 4"); caps = gst_caps_new_simple ("video/x-indeo", "indeoversion", G_TYPE_INT, 4, NULL); break; case GST_MAKE_FOURCC ('d', 'v', 'c', 'p'): case GST_MAKE_FOURCC ('d', 'v', 'c', ' '): case GST_MAKE_FOURCC ('d', 'v', 's', 'd'): case GST_MAKE_FOURCC ('D', 'V', 'S', 'D'): case GST_MAKE_FOURCC ('d', 'v', 'c', 's'): case GST_MAKE_FOURCC ('D', 'V', 'C', 'S'): case GST_MAKE_FOURCC ('d', 'v', '2', '5'): case GST_MAKE_FOURCC ('d', 'v', 'p', 'p'): _codec ("DV Video"); caps = gst_caps_new_simple ("video/x-dv", "dvversion", G_TYPE_INT, 25, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MAKE_FOURCC ('d', 'v', '5', 'n'): //DVCPRO50 NTSC case GST_MAKE_FOURCC ('d', 'v', '5', 'p'): //DVCPRO50 PAL _codec ("DVCPro50 Video"); caps = gst_caps_new_simple ("video/x-dv", "dvversion", G_TYPE_INT, 50, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MAKE_FOURCC ('d', 'v', 'h', '5'): //DVCPRO HD 50i produced by FCP case GST_MAKE_FOURCC ('d', 'v', 'h', '6'): //DVCPRO HD 60i produced by FCP _codec ("DVCProHD Video"); caps = gst_caps_new_simple ("video/x-dv", "dvversion", G_TYPE_INT, 100, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MAKE_FOURCC ('s', 'm', 'c', ' '): _codec ("Apple Graphics (SMC)"); caps = gst_caps_new_simple ("video/x-smc", NULL); break; case GST_MAKE_FOURCC ('V', 'P', '3', '1'): _codec ("VP3"); caps = gst_caps_new_simple ("video/x-vp3", NULL); break; case GST_MAKE_FOURCC ('X', 'i', 'T', 'h'): _codec ("Theora"); caps = gst_caps_new_simple ("video/x-theora", NULL); /* theora uses one byte of padding in the data stream because it does not * allow 0 sized packets while theora does */ stream->padding = 1; break; case GST_MAKE_FOURCC ('d', 'r', 'a', 'c'): _codec ("Dirac"); caps = gst_caps_new_simple ("video/x-dirac", NULL); break; case GST_MAKE_FOURCC ('t', 'i', 'f', 'f'): _codec ("TIFF still images"); caps = gst_caps_new_simple ("image/tiff", NULL); break; case GST_MAKE_FOURCC ('i', 'c', 'o', 'd'): _codec ("Apple Intermediate Codec"); caps = gst_caps_from_string ("video/x-apple-intermediate-codec"); break; case GST_MAKE_FOURCC ('A', 'V', 'd', 'n'): _codec ("AVID DNxHD"); caps = gst_caps_from_string ("video/x-dnxhd"); break; case GST_MAKE_FOURCC ('k', 'p', 'c', 'd'): default: { char *s; s = g_strdup_printf ("video/x-gst-fourcc-%" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (fourcc)); caps = gst_caps_new_simple (s, NULL); break; } } /* enable clipping for raw video streams */ s = gst_caps_get_structure (caps, 0); name = gst_structure_get_name (s); if (g_str_has_prefix (name, "video/x-raw-")) { stream->need_clip = TRUE; } return caps; } static GstCaps * qtdemux_audio_caps (GstQTDemux * qtdemux, QtDemuxStream * stream, guint32 fourcc, const guint8 * data, int len, gchar ** codec_name) { GstCaps *caps; const GstStructure *s; const gchar *name; gint endian = 0; GST_DEBUG_OBJECT (qtdemux, "resolve fourcc %08x", fourcc); switch (fourcc) { case GST_MAKE_FOURCC ('N', 'O', 'N', 'E'): case GST_MAKE_FOURCC ('r', 'a', 'w', ' '): _codec ("Raw 8-bit PCM audio"); caps = gst_caps_new_simple ("audio/x-raw-int", "width", G_TYPE_INT, 8, "depth", G_TYPE_INT, 8, "signed", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MAKE_FOURCC ('t', 'w', 'o', 's'): endian = G_BIG_ENDIAN; /* fall-through */ case GST_MAKE_FOURCC ('s', 'o', 'w', 't'): { gchar *str; gint depth; if (!endian) endian = G_LITTLE_ENDIAN; depth = stream->bytes_per_packet * 8; str = g_strdup_printf ("Raw %d-bit PCM audio", depth); _codec (str); g_free (str); caps = gst_caps_new_simple ("audio/x-raw-int", "width", G_TYPE_INT, depth, "depth", G_TYPE_INT, depth, "endianness", G_TYPE_INT, endian, "signed", G_TYPE_BOOLEAN, TRUE, NULL); break; } case GST_MAKE_FOURCC ('f', 'l', '6', '4'): _codec ("Raw 64-bit floating-point audio"); caps = gst_caps_new_simple ("audio/x-raw-float", "width", G_TYPE_INT, 64, "endianness", G_TYPE_INT, G_BIG_ENDIAN, NULL); break; case GST_MAKE_FOURCC ('f', 'l', '3', '2'): _codec ("Raw 32-bit floating-point audio"); caps = gst_caps_new_simple ("audio/x-raw-float", "width", G_TYPE_INT, 32, "endianness", G_TYPE_INT, G_BIG_ENDIAN, NULL); break; case FOURCC_in24: _codec ("Raw 24-bit PCM audio"); /* we assume BIG ENDIAN, an enda box will tell us to change this to little * endian later */ caps = gst_caps_new_simple ("audio/x-raw-int", "width", G_TYPE_INT, 24, "depth", G_TYPE_INT, 24, "endianness", G_TYPE_INT, G_BIG_ENDIAN, "signed", G_TYPE_BOOLEAN, TRUE, NULL); break; case GST_MAKE_FOURCC ('i', 'n', '3', '2'): _codec ("Raw 32-bit PCM audio"); caps = gst_caps_new_simple ("audio/x-raw-int", "width", G_TYPE_INT, 32, "depth", G_TYPE_INT, 32, "endianness", G_TYPE_INT, G_BIG_ENDIAN, "signed", G_TYPE_BOOLEAN, TRUE, NULL); break; case GST_MAKE_FOURCC ('u', 'l', 'a', 'w'): _codec ("Mu-law audio"); caps = gst_caps_new_simple ("audio/x-mulaw", NULL); break; case GST_MAKE_FOURCC ('a', 'l', 'a', 'w'): _codec ("A-law audio"); caps = gst_caps_new_simple ("audio/x-alaw", NULL); break; case 0x0200736d: case 0x6d730002: _codec ("Microsoft ADPCM"); /* Microsoft ADPCM-ACM code 2 */ caps = gst_caps_new_simple ("audio/x-adpcm", "layout", G_TYPE_STRING, "microsoft", NULL); break; case 0x1100736d: case 0x6d730011: _codec ("IMA Loki SDL MJPEG ADPCM"); /* Loki ADPCM, See #550288 for a file that only decodes * with the smjpeg variant of the ADPCM decoder. */ caps = gst_caps_new_simple ("audio/x-adpcm", "layout", G_TYPE_STRING, "smjpeg", NULL); break; case 0x1700736d: case 0x6d730017: _codec ("DVI/Intel IMA ADPCM"); /* FIXME DVI/Intel IMA ADPCM/ACM code 17 */ caps = gst_caps_new_simple ("audio/x-adpcm", "layout", G_TYPE_STRING, "quicktime", NULL); break; case 0x5500736d: case 0x6d730055: /* MPEG layer 3, CBR only (pre QT4.1) */ case GST_MAKE_FOURCC ('.', 'm', 'p', '3'): _codec ("MPEG-1 layer 3"); /* MPEG layer 3, CBR & VBR (QT4.1 and later) */ caps = gst_caps_new_simple ("audio/mpeg", "layer", G_TYPE_INT, 3, "mpegversion", G_TYPE_INT, 1, NULL); break; case 0x20736d: case GST_MAKE_FOURCC ('a', 'c', '-', '3'): _codec ("AC-3 audio"); caps = gst_caps_new_simple ("audio/x-ac3", NULL); stream->sampled = TRUE; break; case GST_MAKE_FOURCC ('M', 'A', 'C', '3'): _codec ("MACE-3"); caps = gst_caps_new_simple ("audio/x-mace", "maceversion", G_TYPE_INT, 3, NULL); break; case GST_MAKE_FOURCC ('M', 'A', 'C', '6'): _codec ("MACE-6"); caps = gst_caps_new_simple ("audio/x-mace", "maceversion", G_TYPE_INT, 6, NULL); break; case GST_MAKE_FOURCC ('O', 'g', 'g', 'V'): /* ogg/vorbis */ caps = gst_caps_new_simple ("application/ogg", NULL); break; case GST_MAKE_FOURCC ('d', 'v', 'c', 'a'): _codec ("DV audio"); caps = gst_caps_new_simple ("audio/x-dv", NULL); break; case GST_MAKE_FOURCC ('m', 'p', '4', 'a'): _codec ("MPEG-4 AAC audio"); caps = gst_caps_new_simple ("audio/mpeg", "mpegversion", G_TYPE_INT, 4, "framed", G_TYPE_BOOLEAN, TRUE, NULL); break; case GST_MAKE_FOURCC ('Q', 'D', 'M', 'C'): _codec ("QDesign Music"); caps = gst_caps_new_simple ("audio/x-qdm", NULL); break; case GST_MAKE_FOURCC ('Q', 'D', 'M', '2'): _codec ("QDesign Music v.2"); /* FIXME: QDesign music version 2 (no constant) */ if (data) { caps = gst_caps_new_simple ("audio/x-qdm2", "framesize", G_TYPE_INT, QT_UINT32 (data + 52), "bitrate", G_TYPE_INT, QT_UINT32 (data + 40), "blocksize", G_TYPE_INT, QT_UINT32 (data + 44), NULL); } else { caps = gst_caps_new_simple ("audio/x-qdm2", NULL); } break; case GST_MAKE_FOURCC ('a', 'g', 's', 'm'): _codec ("GSM audio"); caps = gst_caps_new_simple ("audio/x-gsm", NULL); break; case GST_MAKE_FOURCC ('s', 'a', 'm', 'r'): _codec ("AMR audio"); caps = gst_caps_new_simple ("audio/AMR", NULL); break; case GST_MAKE_FOURCC ('s', 'a', 'w', 'b'): _codec ("AMR-WB audio"); caps = gst_caps_new_simple ("audio/AMR-WB", NULL); break; case GST_MAKE_FOURCC ('i', 'm', 'a', '4'): _codec ("Quicktime IMA ADPCM"); caps = gst_caps_new_simple ("audio/x-adpcm", "layout", G_TYPE_STRING, "quicktime", NULL); break; case GST_MAKE_FOURCC ('a', 'l', 'a', 'c'): _codec ("Apple lossless audio"); caps = gst_caps_new_simple ("audio/x-alac", NULL); break; case GST_MAKE_FOURCC ('Q', 'c', 'l', 'p'): _codec ("QualComm PureVoice"); caps = gst_caps_from_string ("audio/qcelp"); break; case GST_MAKE_FOURCC ('q', 't', 'v', 'r'): /* ? */ default: { char *s; s = g_strdup_printf ("audio/x-gst-fourcc-%" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (fourcc)); caps = gst_caps_new_simple (s, NULL); break; } } /* enable clipping for raw audio streams */ s = gst_caps_get_structure (caps, 0); name = gst_structure_get_name (s); if (g_str_has_prefix (name, "audio/x-raw-")) { stream->need_clip = TRUE; } return caps; } static GstCaps * qtdemux_sub_caps (GstQTDemux * qtdemux, QtDemuxStream * stream, guint32 fourcc, const guint8 * stsd_data, gchar ** codec_name) { GstCaps *caps; GST_DEBUG_OBJECT (qtdemux, "resolve fourcc %08x", fourcc); switch (fourcc) { case GST_MAKE_FOURCC ('m', 'p', '4', 's'): _codec ("DVD subtitle"); caps = gst_caps_new_simple ("video/x-dvd-subpicture", NULL); break; case GST_MAKE_FOURCC ('t', 'x', '3', 'g'): _codec ("3GPP timed text"); caps = gst_caps_new_simple ("text/plain", NULL); /* actual text piece needs to be extracted */ stream->need_process = TRUE; break; default: { char *s; s = g_strdup_printf ("text/x-gst-fourcc-%" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (fourcc)); caps = gst_caps_new_simple (s, NULL); break; } } return caps; }