| /* |
| * tsdemux.c |
| * Copyright (C) 2009 Zaheer Abbas Merali |
| * 2010 Edward Hervey |
| * Copyright (C) 2011, Hewlett-Packard Development Company, L.P. |
| * Author: Youness Alaoui <youness.alaoui@collabora.co.uk>, Collabora Ltd. |
| * Author: Sebastian Dröge <sebastian.droege@collabora.co.uk>, Collabora Ltd. |
| * Author: Edward Hervey <bilboed@bilboed.com>, Collabora Ltd. |
| * |
| * Authors: |
| * Zaheer Abbas Merali <zaheerabbas at merali dot org> |
| * Edward Hervey <edward.hervey@collabora.co.uk> |
| * |
| * This library is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Library General Public |
| * License as published by the Free Software Foundation; either |
| * version 2 of the License, or (at your option) any later version. |
| * |
| * This library is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Library General Public License for more details. |
| * |
| * You should have received a copy of the GNU Library General Public |
| * License along with this library; if not, write to the |
| * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, |
| * Boston, MA 02110-1301, USA. |
| */ |
| |
| #ifdef HAVE_CONFIG_H |
| #include "config.h" |
| #endif |
| |
| #include <stdlib.h> |
| #include <string.h> |
| |
| #include <glib.h> |
| #include <gst/tag/tag.h> |
| #include <gst/pbutils/pbutils.h> |
| #include <gst/base/base.h> |
| #include <gst/audio/audio.h> |
| |
| #include "mpegtsbase.h" |
| #include "tsdemux.h" |
| #include "gstmpegdesc.h" |
| #include "gstmpegdefs.h" |
| #include "mpegtspacketizer.h" |
| #include "pesparse.h" |
| #include <gst/codecparsers/gsth264parser.h> |
| #include <gst/codecparsers/gstmpegvideoparser.h> |
| #include <gst/video/video-color.h> |
| |
| #include <math.h> |
| |
| #define _gst_log2(x) (log(x)/log(2)) |
| |
| /* |
| * tsdemux |
| * |
| * See TODO for explanations on improvements needed |
| */ |
| |
| #define CONTINUITY_UNSET 255 |
| #define MAX_CONTINUITY 15 |
| |
| /* Seeking/Scanning related variables */ |
| |
| /* seek to SEEK_TIMESTAMP_OFFSET before the desired offset and search then |
| * either accurately or for the next timestamp |
| */ |
| #define SEEK_TIMESTAMP_OFFSET (2500 * GST_MSECOND) |
| |
| #define GST_FLOW_REWINDING GST_FLOW_CUSTOM_ERROR |
| |
| /* latency in nsecs */ |
| #define TS_LATENCY (700 * GST_MSECOND) |
| |
| GST_DEBUG_CATEGORY_STATIC (ts_demux_debug); |
| #define GST_CAT_DEFAULT ts_demux_debug |
| |
| #define ABSDIFF(a,b) (((a) > (b)) ? ((a) - (b)) : ((b) - (a))) |
| |
| static GQuark QUARK_TSDEMUX; |
| static GQuark QUARK_PID; |
| static GQuark QUARK_PCR; |
| static GQuark QUARK_OPCR; |
| static GQuark QUARK_PTS; |
| static GQuark QUARK_DTS; |
| static GQuark QUARK_OFFSET; |
| |
| typedef enum |
| { |
| PENDING_PACKET_EMPTY = 0, /* No pending packet/buffer |
| * Push incoming buffers to the array */ |
| PENDING_PACKET_HEADER, /* PES header needs to be parsed |
| * Push incoming buffers to the array */ |
| PENDING_PACKET_BUFFER, /* Currently filling up output buffer |
| * Push incoming buffers to the bufferlist */ |
| PENDING_PACKET_DISCONT /* Discontinuity in incoming packets |
| * Drop all incoming buffers */ |
| } PendingPacketState; |
| |
| /* Pending buffer */ |
| typedef struct |
| { |
| /* The fully reconstructed buffer */ |
| GstBuffer *buffer; |
| |
| /* Raw PTS/DTS (in 90kHz units) */ |
| guint64 pts, dts; |
| } PendingBuffer; |
| |
| typedef struct _TSDemuxStream TSDemuxStream; |
| |
| typedef struct _TSDemuxH264ParsingInfos TSDemuxH264ParsingInfos; |
| typedef struct _TSDemuxJP2KParsingInfos TSDemuxJP2KParsingInfos; |
| |
| /* Returns TRUE if a keyframe was found */ |
| typedef gboolean (*GstTsDemuxKeyFrameScanFunction) (TSDemuxStream * stream, |
| guint8 * data, const gsize data_size, const gsize max_frame_offset); |
| |
| typedef struct |
| { |
| guint8 *data; |
| gsize size; |
| } SimpleBuffer; |
| |
| struct _TSDemuxH264ParsingInfos |
| { |
| /* H264 parsing data */ |
| GstH264NalParser *parser; |
| GstByteWriter *sps; |
| GstByteWriter *pps; |
| GstByteWriter *sei; |
| SimpleBuffer framedata; |
| }; |
| |
| struct _TSDemuxJP2KParsingInfos |
| { |
| /* J2K parsing data */ |
| gboolean interlace; |
| }; |
| struct _TSDemuxStream |
| { |
| MpegTSBaseStream stream; |
| |
| GstPad *pad; |
| |
| /* Whether the pad was added or not */ |
| gboolean active; |
| |
| /* Whether this is a sparse stream (subtitles or metadata) */ |
| gboolean sparse; |
| |
| /* TRUE if we are waiting for a valid timestamp */ |
| gboolean pending_ts; |
| |
| /* Output data */ |
| PendingPacketState state; |
| |
| /* Data being reconstructed (allocated) */ |
| guint8 *data; |
| |
| /* Size of data being reconstructed (if known, else 0) */ |
| guint expected_size; |
| |
| /* Amount of bytes in current ->data */ |
| guint current_size; |
| /* Size of ->data */ |
| guint allocated_size; |
| |
| /* Current PTS/DTS for this stream (in running time) */ |
| GstClockTime pts; |
| GstClockTime dts; |
| |
| /* Reference PTS used to detect gaps */ |
| GstClockTime gap_ref_pts; |
| /* Number of outputted buffers */ |
| guint32 nb_out_buffers; |
| /* Reference number of buffers for gaps */ |
| guint32 gap_ref_buffers; |
| |
| /* Current PTS/DTS for this stream (in 90kHz unit) */ |
| guint64 raw_pts, raw_dts; |
| |
| /* Whether this stream needs to send a newsegment */ |
| gboolean need_newsegment; |
| |
| /* Whether the next output buffer should be DISCONT */ |
| gboolean discont; |
| |
| /* The value to use when calculating the newsegment */ |
| GstClockTime first_pts; |
| |
| GstTagList *taglist; |
| |
| gint continuity_counter; |
| |
| /* List of pending buffers */ |
| GList *pending; |
| |
| /* if != 0, output only PES from that substream */ |
| guint8 target_pes_substream; |
| gboolean needs_keyframe; |
| |
| GstClockTime seeked_pts, seeked_dts; |
| |
| GstTsDemuxKeyFrameScanFunction scan_function; |
| TSDemuxH264ParsingInfos h264infos; |
| TSDemuxJP2KParsingInfos jp2kInfos; |
| }; |
| |
| #define VIDEO_CAPS \ |
| GST_STATIC_CAPS (\ |
| "video/mpeg, " \ |
| "mpegversion = (int) { 1, 2, 4 }, " \ |
| "systemstream = (boolean) FALSE; " \ |
| "video/x-h264,stream-format=(string)byte-stream," \ |
| "alignment=(string)nal;" \ |
| "video/x-h265,stream-format=(string)byte-stream," \ |
| "alignment=(string)nal;" \ |
| "video/x-dirac;" \ |
| "video/x-cavs;" \ |
| "video/x-wmv," \ |
| "wmvversion = (int) 3, " \ |
| "format = (string) WVC1;" \ |
| "image/x-jpc;" \ |
| ) |
| |
| #define AUDIO_CAPS \ |
| GST_STATIC_CAPS ( \ |
| "audio/mpeg, " \ |
| "mpegversion = (int) 1;" \ |
| "audio/mpeg, " \ |
| "mpegversion = (int) 2, " \ |
| "stream-format = (string) adts; " \ |
| "audio/mpeg, " \ |
| "mpegversion = (int) 4, " \ |
| "stream-format = (string) loas; " \ |
| "audio/x-lpcm, " \ |
| "width = (int) { 16, 20, 24 }, " \ |
| "rate = (int) { 48000, 96000 }, " \ |
| "channels = (int) [ 1, 8 ], " \ |
| "dynamic_range = (int) [ 0, 255 ], " \ |
| "emphasis = (boolean) { FALSE, TRUE }, " \ |
| "mute = (boolean) { FALSE, TRUE }; " \ |
| "audio/x-ac3; audio/x-eac3;" \ |
| "audio/x-dts;" \ |
| "audio/x-opus;" \ |
| "audio/x-private-ts-lpcm" \ |
| ) |
| |
| /* Can also use the subpicture pads for text subtitles? */ |
| #define SUBPICTURE_CAPS \ |
| GST_STATIC_CAPS ("subpicture/x-pgs; subpicture/x-dvd; subpicture/x-dvb") |
| |
| static GstStaticPadTemplate video_template = |
| GST_STATIC_PAD_TEMPLATE ("video_%01x_%05x", GST_PAD_SRC, |
| GST_PAD_SOMETIMES, |
| VIDEO_CAPS); |
| |
| static GstStaticPadTemplate audio_template = |
| GST_STATIC_PAD_TEMPLATE ("audio_%01x_%05x", |
| GST_PAD_SRC, |
| GST_PAD_SOMETIMES, |
| AUDIO_CAPS); |
| |
| static GstStaticPadTemplate subpicture_template = |
| GST_STATIC_PAD_TEMPLATE ("subpicture_%01x_%05x", |
| GST_PAD_SRC, |
| GST_PAD_SOMETIMES, |
| SUBPICTURE_CAPS); |
| |
| static GstStaticPadTemplate private_template = |
| GST_STATIC_PAD_TEMPLATE ("private_%01x_%05x", |
| GST_PAD_SRC, |
| GST_PAD_SOMETIMES, |
| GST_STATIC_CAPS_ANY); |
| |
| enum |
| { |
| PROP_0, |
| PROP_PROGRAM_NUMBER, |
| PROP_EMIT_STATS, |
| /* FILL ME */ |
| }; |
| |
| /* Pad functions */ |
| |
| |
| /* mpegtsbase methods */ |
| static void |
| gst_ts_demux_update_program (MpegTSBase * base, MpegTSBaseProgram * program); |
| static void |
| gst_ts_demux_program_started (MpegTSBase * base, MpegTSBaseProgram * program); |
| static void |
| gst_ts_demux_program_stopped (MpegTSBase * base, MpegTSBaseProgram * program); |
| static gboolean |
| gst_ts_demux_can_remove_program (MpegTSBase * base, |
| MpegTSBaseProgram * program); |
| static void gst_ts_demux_reset (MpegTSBase * base); |
| static GstFlowReturn |
| gst_ts_demux_push (MpegTSBase * base, MpegTSPacketizerPacket * packet, |
| GstMpegtsSection * section); |
| static void gst_ts_demux_flush (MpegTSBase * base, gboolean hard); |
| static GstFlowReturn gst_ts_demux_drain (MpegTSBase * base); |
| static gboolean |
| gst_ts_demux_stream_added (MpegTSBase * base, MpegTSBaseStream * stream, |
| MpegTSBaseProgram * program); |
| static void |
| gst_ts_demux_stream_removed (MpegTSBase * base, MpegTSBaseStream * stream); |
| static GstFlowReturn gst_ts_demux_do_seek (MpegTSBase * base, GstEvent * event); |
| static void gst_ts_demux_set_property (GObject * object, guint prop_id, |
| const GValue * value, GParamSpec * pspec); |
| static void gst_ts_demux_get_property (GObject * object, guint prop_id, |
| GValue * value, GParamSpec * pspec); |
| static void gst_ts_demux_flush_streams (GstTSDemux * tsdemux, gboolean hard); |
| static GstFlowReturn |
| gst_ts_demux_push_pending_data (GstTSDemux * demux, TSDemuxStream * stream, |
| MpegTSBaseProgram * program); |
| static void gst_ts_demux_stream_flush (TSDemuxStream * stream, |
| GstTSDemux * demux, gboolean hard); |
| |
| static gboolean push_event (MpegTSBase * base, GstEvent * event); |
| static void gst_ts_demux_check_and_sync_streams (GstTSDemux * demux, |
| GstClockTime time); |
| |
| static void |
| _extra_init (void) |
| { |
| QUARK_TSDEMUX = g_quark_from_string ("tsdemux"); |
| QUARK_PID = g_quark_from_string ("pid"); |
| QUARK_PCR = g_quark_from_string ("pcr"); |
| QUARK_OPCR = g_quark_from_string ("opcr"); |
| QUARK_PTS = g_quark_from_string ("pts"); |
| QUARK_DTS = g_quark_from_string ("dts"); |
| QUARK_OFFSET = g_quark_from_string ("offset"); |
| } |
| |
| #define gst_ts_demux_parent_class parent_class |
| G_DEFINE_TYPE_WITH_CODE (GstTSDemux, gst_ts_demux, GST_TYPE_MPEGTS_BASE, |
| _extra_init ()); |
| |
| static void |
| gst_ts_demux_dispose (GObject * object) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX_CAST (object); |
| |
| gst_flow_combiner_free (demux->flowcombiner); |
| |
| GST_CALL_PARENT (G_OBJECT_CLASS, dispose, (object)); |
| } |
| |
| static void |
| gst_ts_demux_class_init (GstTSDemuxClass * klass) |
| { |
| GObjectClass *gobject_class; |
| GstElementClass *element_class; |
| MpegTSBaseClass *ts_class; |
| |
| gobject_class = G_OBJECT_CLASS (klass); |
| gobject_class->set_property = gst_ts_demux_set_property; |
| gobject_class->get_property = gst_ts_demux_get_property; |
| gobject_class->dispose = gst_ts_demux_dispose; |
| |
| g_object_class_install_property (gobject_class, PROP_PROGRAM_NUMBER, |
| g_param_spec_int ("program-number", "Program number", |
| "Program Number to demux for (-1 to ignore)", -1, G_MAXINT, |
| -1, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); |
| |
| g_object_class_install_property (gobject_class, PROP_EMIT_STATS, |
| g_param_spec_boolean ("emit-stats", "Emit statistics", |
| "Emit messages for every pcr/opcr/pts/dts", FALSE, |
| G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); |
| |
| element_class = GST_ELEMENT_CLASS (klass); |
| gst_element_class_add_pad_template (element_class, |
| gst_static_pad_template_get (&video_template)); |
| gst_element_class_add_pad_template (element_class, |
| gst_static_pad_template_get (&audio_template)); |
| gst_element_class_add_pad_template (element_class, |
| gst_static_pad_template_get (&subpicture_template)); |
| gst_element_class_add_pad_template (element_class, |
| gst_static_pad_template_get (&private_template)); |
| |
| gst_element_class_set_static_metadata (element_class, |
| "MPEG transport stream demuxer", |
| "Codec/Demuxer", |
| "Demuxes MPEG2 transport streams", |
| "Zaheer Abbas Merali <zaheerabbas at merali dot org>\n" |
| "Edward Hervey <edward.hervey@collabora.co.uk>"); |
| |
| ts_class = GST_MPEGTS_BASE_CLASS (klass); |
| ts_class->reset = GST_DEBUG_FUNCPTR (gst_ts_demux_reset); |
| ts_class->push = GST_DEBUG_FUNCPTR (gst_ts_demux_push); |
| ts_class->push_event = GST_DEBUG_FUNCPTR (push_event); |
| ts_class->program_started = GST_DEBUG_FUNCPTR (gst_ts_demux_program_started); |
| ts_class->program_stopped = GST_DEBUG_FUNCPTR (gst_ts_demux_program_stopped); |
| ts_class->update_program = GST_DEBUG_FUNCPTR (gst_ts_demux_update_program); |
| ts_class->can_remove_program = gst_ts_demux_can_remove_program; |
| ts_class->stream_added = gst_ts_demux_stream_added; |
| ts_class->stream_removed = gst_ts_demux_stream_removed; |
| ts_class->seek = GST_DEBUG_FUNCPTR (gst_ts_demux_do_seek); |
| ts_class->flush = GST_DEBUG_FUNCPTR (gst_ts_demux_flush); |
| ts_class->drain = GST_DEBUG_FUNCPTR (gst_ts_demux_drain); |
| } |
| |
| static void |
| gst_ts_demux_reset (MpegTSBase * base) |
| { |
| GstTSDemux *demux = (GstTSDemux *) base; |
| |
| demux->rate = 1.0; |
| gst_segment_init (&demux->segment, GST_FORMAT_UNDEFINED); |
| if (demux->segment_event) { |
| gst_event_unref (demux->segment_event); |
| demux->segment_event = NULL; |
| } |
| |
| if (demux->global_tags) { |
| gst_tag_list_unref (demux->global_tags); |
| demux->global_tags = NULL; |
| } |
| |
| if (demux->previous_program) { |
| mpegts_base_deactivate_and_free_program (base, demux->previous_program); |
| demux->previous_program = NULL; |
| } |
| |
| demux->have_group_id = FALSE; |
| demux->group_id = G_MAXUINT; |
| |
| demux->last_seek_offset = -1; |
| demux->program_generation = 0; |
| } |
| |
| static void |
| gst_ts_demux_init (GstTSDemux * demux) |
| { |
| MpegTSBase *base = (MpegTSBase *) demux; |
| |
| base->stream_size = sizeof (TSDemuxStream); |
| base->parse_private_sections = TRUE; |
| /* We are not interested in sections (all handled by mpegtsbase) */ |
| base->push_section = FALSE; |
| |
| demux->flowcombiner = gst_flow_combiner_new (); |
| demux->requested_program_number = -1; |
| demux->program_number = -1; |
| gst_ts_demux_reset (base); |
| } |
| |
| |
| static void |
| gst_ts_demux_set_property (GObject * object, guint prop_id, |
| const GValue * value, GParamSpec * pspec) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX (object); |
| |
| switch (prop_id) { |
| case PROP_PROGRAM_NUMBER: |
| /* FIXME: do something if program is switched as opposed to set at |
| * beginning */ |
| demux->requested_program_number = g_value_get_int (value); |
| break; |
| case PROP_EMIT_STATS: |
| demux->emit_statistics = g_value_get_boolean (value); |
| break; |
| default: |
| G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); |
| } |
| } |
| |
| static void |
| gst_ts_demux_get_property (GObject * object, guint prop_id, |
| GValue * value, GParamSpec * pspec) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX (object); |
| |
| switch (prop_id) { |
| case PROP_PROGRAM_NUMBER: |
| g_value_set_int (value, demux->requested_program_number); |
| break; |
| case PROP_EMIT_STATS: |
| g_value_set_boolean (value, demux->emit_statistics); |
| break; |
| default: |
| G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); |
| } |
| } |
| |
| static gboolean |
| gst_ts_demux_get_duration (GstTSDemux * demux, GstClockTime * dur) |
| { |
| MpegTSBase *base = (MpegTSBase *) demux; |
| gboolean res = FALSE; |
| gint64 val; |
| |
| /* Get total size in bytes */ |
| if (gst_pad_peer_query_duration (base->sinkpad, GST_FORMAT_BYTES, &val)) { |
| /* Convert it to duration */ |
| *dur = |
| mpegts_packetizer_offset_to_ts (base->packetizer, val, |
| demux->program->pcr_pid); |
| if (GST_CLOCK_TIME_IS_VALID (*dur)) |
| res = TRUE; |
| } |
| return res; |
| } |
| |
| static gboolean |
| gst_ts_demux_srcpad_query (GstPad * pad, GstObject * parent, GstQuery * query) |
| { |
| gboolean res = TRUE; |
| GstFormat format; |
| GstTSDemux *demux; |
| MpegTSBase *base; |
| |
| demux = GST_TS_DEMUX (parent); |
| base = GST_MPEGTS_BASE (demux); |
| |
| switch (GST_QUERY_TYPE (query)) { |
| case GST_QUERY_DURATION: |
| { |
| GST_DEBUG ("query duration"); |
| gst_query_parse_duration (query, &format, NULL); |
| if (format == GST_FORMAT_TIME) { |
| if (!gst_pad_peer_query (base->sinkpad, query)) { |
| GstClockTime dur; |
| if (gst_ts_demux_get_duration (demux, &dur)) |
| gst_query_set_duration (query, GST_FORMAT_TIME, dur); |
| else |
| res = FALSE; |
| } |
| } else { |
| GST_DEBUG_OBJECT (demux, "only query duration on TIME is supported"); |
| res = FALSE; |
| } |
| break; |
| } |
| case GST_QUERY_LATENCY: |
| { |
| GST_DEBUG ("query latency"); |
| res = gst_pad_peer_query (base->sinkpad, query); |
| if (res) { |
| GstClockTime min_lat, max_lat; |
| gboolean live; |
| |
| /* According to H.222.0 |
| Annex D.0.3 (System Time Clock recovery in the decoder) |
| and D.0.2 (Audio and video presentation synchronization) |
| |
| We can end up with an interval of up to 700ms between valid |
| PTS/DTS. We therefore allow a latency of 700ms for that. |
| */ |
| gst_query_parse_latency (query, &live, &min_lat, &max_lat); |
| min_lat += TS_LATENCY; |
| if (GST_CLOCK_TIME_IS_VALID (max_lat)) |
| max_lat += TS_LATENCY; |
| gst_query_set_latency (query, live, min_lat, max_lat); |
| } |
| break; |
| } |
| case GST_QUERY_SEEKING: |
| { |
| GST_DEBUG ("query seeking"); |
| gst_query_parse_seeking (query, &format, NULL, NULL, NULL); |
| GST_DEBUG ("asked for format %s", gst_format_get_name (format)); |
| if (format == GST_FORMAT_TIME) { |
| gboolean seekable = FALSE; |
| |
| if (gst_pad_peer_query (base->sinkpad, query)) |
| gst_query_parse_seeking (query, NULL, &seekable, NULL, NULL); |
| |
| /* If upstream is not seekable in TIME format we use |
| * our own values here */ |
| if (!seekable) { |
| GstClockTime dur; |
| if (gst_ts_demux_get_duration (demux, &dur)) { |
| gst_query_set_seeking (query, GST_FORMAT_TIME, TRUE, 0, dur); |
| GST_DEBUG ("Gave duration: %" GST_TIME_FORMAT, GST_TIME_ARGS (dur)); |
| } |
| } |
| } else { |
| GST_DEBUG_OBJECT (demux, "only TIME is supported for query seeking"); |
| res = FALSE; |
| } |
| break; |
| } |
| case GST_QUERY_SEGMENT:{ |
| GstFormat format; |
| gint64 start, stop; |
| |
| format = demux->segment.format; |
| |
| start = |
| gst_segment_to_stream_time (&demux->segment, format, |
| demux->segment.start); |
| if ((stop = demux->segment.stop) == -1) |
| stop = demux->segment.duration; |
| else |
| stop = gst_segment_to_stream_time (&demux->segment, format, stop); |
| |
| gst_query_set_segment (query, demux->segment.rate, format, start, stop); |
| res = TRUE; |
| break; |
| } |
| default: |
| res = gst_pad_query_default (pad, parent, query); |
| } |
| |
| return res; |
| |
| } |
| |
| static void |
| clear_simple_buffer (SimpleBuffer * sbuf) |
| { |
| if (!sbuf->data) |
| return; |
| |
| g_free (sbuf->data); |
| sbuf->size = 0; |
| sbuf->data = NULL; |
| } |
| |
| static gboolean |
| scan_keyframe_h264 (TSDemuxStream * stream, const guint8 * data, |
| const gsize data_size, const gsize max_frame_offset) |
| { |
| gint offset = 0; |
| GstH264NalUnit unit, frame_unit = { 0, }; |
| GstH264ParserResult res = GST_H264_PARSER_OK; |
| TSDemuxH264ParsingInfos *h264infos = &stream->h264infos; |
| |
| GstH264NalParser *parser = h264infos->parser; |
| |
| if (G_UNLIKELY (parser == NULL)) { |
| parser = h264infos->parser = gst_h264_nal_parser_new (); |
| h264infos->sps = gst_byte_writer_new (); |
| h264infos->pps = gst_byte_writer_new (); |
| h264infos->sei = gst_byte_writer_new (); |
| } |
| |
| while (res == GST_H264_PARSER_OK) { |
| res = |
| gst_h264_parser_identify_nalu (parser, data, offset, data_size, &unit); |
| |
| if (res != GST_H264_PARSER_OK && res != GST_H264_PARSER_NO_NAL_END) { |
| GST_INFO_OBJECT (stream->pad, "Error identifying nalu: %i", res); |
| break; |
| } |
| |
| res = gst_h264_parser_parse_nal (parser, &unit); |
| if (res != GST_H264_PARSER_OK) { |
| break; |
| } |
| |
| switch (unit.type) { |
| case GST_H264_NAL_SEI: |
| if (frame_unit.size) |
| break; |
| |
| if (gst_byte_writer_put_data (h264infos->sei, |
| unit.data + unit.sc_offset, |
| unit.size + unit.offset - unit.sc_offset)) { |
| GST_DEBUG ("adding SEI %u", unit.size + unit.offset - unit.sc_offset); |
| } else { |
| GST_WARNING ("Could not write SEI"); |
| } |
| break; |
| case GST_H264_NAL_PPS: |
| if (frame_unit.size) |
| break; |
| |
| if (gst_byte_writer_put_data (h264infos->pps, |
| unit.data + unit.sc_offset, |
| unit.size + unit.offset - unit.sc_offset)) { |
| GST_DEBUG ("adding PPS %u", unit.size + unit.offset - unit.sc_offset); |
| } else { |
| GST_WARNING ("Could not write PPS"); |
| } |
| break; |
| case GST_H264_NAL_SPS: |
| if (frame_unit.size) |
| break; |
| |
| if (gst_byte_writer_put_data (h264infos->sps, |
| unit.data + unit.sc_offset, |
| unit.size + unit.offset - unit.sc_offset)) { |
| GST_DEBUG ("adding SPS %u", unit.size + unit.offset - unit.sc_offset); |
| } else { |
| GST_WARNING ("Could not write SPS"); |
| } |
| break; |
| /* these units are considered keyframes in h264parse */ |
| case GST_H264_NAL_SLICE: |
| case GST_H264_NAL_SLICE_DPA: |
| case GST_H264_NAL_SLICE_DPB: |
| case GST_H264_NAL_SLICE_DPC: |
| case GST_H264_NAL_SLICE_IDR: |
| { |
| GstH264SliceHdr slice; |
| |
| if (h264infos->framedata.size) |
| break; |
| |
| res = gst_h264_parser_parse_slice_hdr (parser, &unit, &slice, |
| FALSE, FALSE); |
| |
| if (GST_H264_IS_I_SLICE (&slice) || GST_H264_IS_SI_SLICE (&slice)) { |
| if (*(unit.data + unit.offset + 1) & 0x80) { |
| /* means first_mb_in_slice == 0 */ |
| /* real frame data */ |
| GST_DEBUG_OBJECT (stream->pad, "Found keyframe at: %u", |
| unit.sc_offset); |
| frame_unit = unit; |
| } |
| } |
| |
| break; |
| } |
| default: |
| break; |
| } |
| |
| if (offset == unit.sc_offset + unit.size) |
| break; |
| |
| offset = unit.sc_offset + unit.size; |
| } |
| |
| /* We've got all the infos we need (SPS / PPS and a keyframe, plus |
| * and possibly SEI units. We can stop rewinding the stream |
| */ |
| if (gst_byte_writer_get_size (h264infos->sps) && |
| gst_byte_writer_get_size (h264infos->pps) && |
| (h264infos->framedata.size || frame_unit.size)) { |
| guint8 *data = NULL; |
| |
| gsize tmpsize = gst_byte_writer_get_size (h264infos->pps); |
| |
| /* We know that the SPS is first so just put all our data in there */ |
| data = gst_byte_writer_reset_and_get_data (h264infos->pps); |
| gst_byte_writer_put_data (h264infos->sps, data, tmpsize); |
| g_free (data); |
| |
| tmpsize = gst_byte_writer_get_size (h264infos->sei); |
| if (tmpsize) { |
| GST_DEBUG ("Adding SEI"); |
| data = gst_byte_writer_reset_and_get_data (h264infos->sei); |
| gst_byte_writer_put_data (h264infos->sps, data, tmpsize); |
| g_free (data); |
| } |
| |
| if (frame_unit.size) { /* We found the everything in one go! */ |
| GST_DEBUG ("Adding Keyframe"); |
| gst_byte_writer_put_data (h264infos->sps, |
| frame_unit.data + frame_unit.sc_offset, |
| stream->current_size - frame_unit.sc_offset); |
| } else { |
| GST_DEBUG ("Adding Keyframe"); |
| gst_byte_writer_put_data (h264infos->sps, |
| h264infos->framedata.data, h264infos->framedata.size); |
| clear_simple_buffer (&h264infos->framedata); |
| } |
| |
| g_free (stream->data); |
| stream->current_size = gst_byte_writer_get_size (h264infos->sps); |
| stream->data = gst_byte_writer_reset_and_get_data (h264infos->sps); |
| gst_byte_writer_init (h264infos->sps); |
| gst_byte_writer_init (h264infos->pps); |
| gst_byte_writer_init (h264infos->sei); |
| |
| return TRUE; |
| } |
| |
| if (frame_unit.size) { |
| GST_DEBUG_OBJECT (stream->pad, "Keep the keyframe as this is the one" |
| " we will push later"); |
| |
| h264infos->framedata.data = |
| g_memdup (frame_unit.data + frame_unit.sc_offset, |
| stream->current_size - frame_unit.sc_offset); |
| h264infos->framedata.size = stream->current_size - frame_unit.sc_offset; |
| } |
| |
| return FALSE; |
| } |
| |
| /* We merge data from TS packets so that the scanning methods get a continuous chunk, |
| however the scanning method will return keyframe offset which needs to be translated |
| back to actual offset in file */ |
| typedef struct |
| { |
| gint64 real_offset; /* offset of TS packet */ |
| gint merged_offset; /* offset of merged data in buffer */ |
| } OffsetInfo; |
| |
| static gboolean |
| gst_ts_demux_adjust_seek_offset_for_keyframe (TSDemuxStream * stream, |
| guint8 * data, guint64 size) |
| { |
| int scan_pid = -1; |
| |
| if (!stream->scan_function) |
| return TRUE; |
| |
| scan_pid = ((MpegTSBaseStream *) stream)->pid; |
| |
| if (scan_pid != -1) { |
| return stream->scan_function (stream, data, size, size); |
| } |
| |
| return TRUE; |
| } |
| |
| static GstFlowReturn |
| gst_ts_demux_do_seek (MpegTSBase * base, GstEvent * event) |
| { |
| GList *tmp; |
| |
| GstTSDemux *demux = (GstTSDemux *) base; |
| GstFlowReturn res = GST_FLOW_ERROR; |
| gdouble rate; |
| GstFormat format; |
| GstSeekFlags flags; |
| GstSeekType start_type, stop_type; |
| gint64 start, stop; |
| guint64 start_offset; |
| |
| gst_event_parse_seek (event, &rate, &format, &flags, &start_type, &start, |
| &stop_type, &stop); |
| |
| GST_DEBUG ("seek event, rate: %f start: %" GST_TIME_FORMAT |
| " stop: %" GST_TIME_FORMAT, rate, GST_TIME_ARGS (start), |
| GST_TIME_ARGS (stop)); |
| |
| if (rate <= 0.0) { |
| GST_WARNING ("Negative rate not supported"); |
| goto done; |
| } |
| |
| if (flags & (GST_SEEK_FLAG_SEGMENT)) { |
| GST_WARNING ("seek flags 0x%x are not supported", (int) flags); |
| goto done; |
| } |
| |
| /* configure the segment with the seek variables */ |
| GST_DEBUG_OBJECT (demux, "configuring seek"); |
| |
| if (start_type != GST_SEEK_TYPE_NONE) { |
| start_offset = |
| mpegts_packetizer_ts_to_offset (base->packetizer, MAX (0, |
| start - SEEK_TIMESTAMP_OFFSET), demux->program->pcr_pid); |
| |
| if (G_UNLIKELY (start_offset == -1)) { |
| GST_WARNING ("Couldn't convert start position to an offset"); |
| goto done; |
| } |
| } else { |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = tmp->data; |
| |
| stream->need_newsegment = TRUE; |
| } |
| gst_segment_init (&demux->segment, GST_FORMAT_UNDEFINED); |
| if (demux->segment_event) { |
| gst_event_unref (demux->segment_event); |
| demux->segment_event = NULL; |
| } |
| demux->rate = rate; |
| res = GST_FLOW_OK; |
| goto done; |
| } |
| |
| /* record offset and rate */ |
| base->seek_offset = start_offset; |
| demux->last_seek_offset = base->seek_offset; |
| demux->rate = rate; |
| res = GST_FLOW_OK; |
| |
| gst_segment_do_seek (&demux->segment, rate, format, flags, start_type, |
| start, stop_type, stop, NULL); |
| /* Reset segment if we're not doing an accurate seek */ |
| demux->reset_segment = (!(flags & GST_SEEK_FLAG_ACCURATE)); |
| |
| if (demux->segment_event) { |
| gst_event_unref (demux->segment_event); |
| demux->segment_event = NULL; |
| } |
| |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = tmp->data; |
| |
| if (flags & GST_SEEK_FLAG_ACCURATE) |
| stream->needs_keyframe = TRUE; |
| |
| stream->seeked_pts = GST_CLOCK_TIME_NONE; |
| stream->seeked_dts = GST_CLOCK_TIME_NONE; |
| stream->need_newsegment = TRUE; |
| stream->first_pts = GST_CLOCK_TIME_NONE; |
| } |
| |
| done: |
| return res; |
| } |
| |
| static gboolean |
| gst_ts_demux_srcpad_event (GstPad * pad, GstObject * parent, GstEvent * event) |
| { |
| gboolean res = TRUE; |
| GstTSDemux *demux = GST_TS_DEMUX (parent); |
| |
| GST_DEBUG_OBJECT (pad, "Got event %s", |
| gst_event_type_get_name (GST_EVENT_TYPE (event))); |
| |
| switch (GST_EVENT_TYPE (event)) { |
| case GST_EVENT_SEEK: |
| res = mpegts_base_handle_seek_event ((MpegTSBase *) demux, pad, event); |
| if (!res) |
| GST_WARNING ("seeking failed"); |
| gst_event_unref (event); |
| break; |
| default: |
| res = gst_pad_event_default (pad, parent, event); |
| } |
| |
| return res; |
| } |
| |
| static void |
| clean_global_taglist (GstTagList * taglist) |
| { |
| gst_tag_list_remove_tag (taglist, GST_TAG_CONTAINER_FORMAT); |
| gst_tag_list_remove_tag (taglist, GST_TAG_CODEC); |
| } |
| |
| static gboolean |
| push_event (MpegTSBase * base, GstEvent * event) |
| { |
| GstTSDemux *demux = (GstTSDemux *) base; |
| GList *tmp; |
| gboolean early_ret = FALSE; |
| |
| if (GST_EVENT_TYPE (event) == GST_EVENT_SEGMENT) { |
| GST_DEBUG_OBJECT (base, "Ignoring segment event (recreated later)"); |
| gst_event_unref (event); |
| return TRUE; |
| |
| } else if (GST_EVENT_TYPE (event) == GST_EVENT_TAG) { |
| /* In case we receive tags before data, store them to send later |
| * If we already have the program, send it right away */ |
| GstTagList *taglist; |
| |
| gst_event_parse_tag (event, &taglist); |
| |
| if (demux->global_tags == NULL) { |
| demux->global_tags = gst_tag_list_copy (taglist); |
| |
| /* Tags that are stream specific for the container should be considered |
| * global for the container streams */ |
| if (gst_tag_list_get_scope (taglist) == GST_TAG_SCOPE_STREAM) { |
| gst_tag_list_set_scope (demux->global_tags, GST_TAG_SCOPE_GLOBAL); |
| } |
| } else { |
| demux->global_tags = gst_tag_list_make_writable (demux->global_tags); |
| gst_tag_list_insert (demux->global_tags, taglist, GST_TAG_MERGE_REPLACE); |
| } |
| clean_global_taglist (demux->global_tags); |
| |
| /* tags are stored to be used after if there are no streams yet, |
| * so we should never reject */ |
| early_ret = TRUE; |
| } |
| |
| if (G_UNLIKELY (demux->program == NULL)) { |
| gst_event_unref (event); |
| return early_ret; |
| } |
| |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = (TSDemuxStream *) tmp->data; |
| if (stream->pad) { |
| /* If we are pushing out EOS, flush out pending data first */ |
| if (GST_EVENT_TYPE (event) == GST_EVENT_EOS && |
| gst_pad_is_active (stream->pad)) |
| gst_ts_demux_push_pending_data (demux, stream, NULL); |
| |
| gst_event_ref (event); |
| gst_pad_push_event (stream->pad, event); |
| } |
| } |
| |
| gst_event_unref (event); |
| |
| return TRUE; |
| } |
| |
| static inline void |
| add_iso639_language_to_tags (TSDemuxStream * stream, gchar * lang_code) |
| { |
| const gchar *lc; |
| |
| GST_LOG ("Add language code for stream: '%s'", lang_code); |
| |
| if (!stream->taglist) |
| stream->taglist = gst_tag_list_new_empty (); |
| |
| /* descriptor contains ISO 639-2 code, we want the ISO 639-1 code */ |
| lc = gst_tag_get_language_code (lang_code); |
| |
| /* Only set tag if we have a valid one */ |
| if (lc || (lang_code[0] && lang_code[1])) |
| gst_tag_list_add (stream->taglist, GST_TAG_MERGE_REPLACE, |
| GST_TAG_LANGUAGE_CODE, (lc) ? lc : lang_code, NULL); |
| } |
| |
| static void |
| gst_ts_demux_create_tags (TSDemuxStream * stream) |
| { |
| MpegTSBaseStream *bstream = (MpegTSBaseStream *) stream; |
| const GstMpegtsDescriptor *desc = NULL; |
| int i, nb; |
| |
| desc = |
| mpegts_get_descriptor_from_stream (bstream, |
| GST_MTS_DESC_ISO_639_LANGUAGE); |
| if (desc) { |
| gchar *lang_code; |
| |
| nb = gst_mpegts_descriptor_parse_iso_639_language_nb (desc); |
| |
| GST_DEBUG ("Found ISO 639 descriptor (%d entries)", nb); |
| |
| for (i = 0; i < nb; i++) |
| if (gst_mpegts_descriptor_parse_iso_639_language_idx (desc, i, &lang_code, |
| NULL)) { |
| add_iso639_language_to_tags (stream, lang_code); |
| g_free (lang_code); |
| } |
| |
| return; |
| } |
| |
| desc = |
| mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_SUBTITLING); |
| |
| if (desc) { |
| gchar *lang_code; |
| |
| nb = gst_mpegts_descriptor_parse_dvb_subtitling_nb (desc); |
| |
| GST_DEBUG ("Found SUBTITLING descriptor (%d entries)", nb); |
| |
| for (i = 0; i < nb; i++) |
| if (gst_mpegts_descriptor_parse_dvb_subtitling_idx (desc, i, &lang_code, |
| NULL, NULL, NULL)) { |
| add_iso639_language_to_tags (stream, lang_code); |
| g_free (lang_code); |
| } |
| } |
| } |
| |
| static GstPad * |
| create_pad_for_stream (MpegTSBase * base, MpegTSBaseStream * bstream, |
| MpegTSBaseProgram * program) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX (base); |
| TSDemuxStream *stream = (TSDemuxStream *) bstream; |
| gchar *name = NULL; |
| GstCaps *caps = NULL; |
| GstPadTemplate *template = NULL; |
| const GstMpegtsDescriptor *desc = NULL; |
| GstPad *pad = NULL; |
| gboolean sparse = FALSE; |
| gboolean is_audio = FALSE, is_video = FALSE, is_subpicture = FALSE, |
| is_private = FALSE; |
| |
| gst_ts_demux_create_tags (stream); |
| |
| GST_LOG ("Attempting to create pad for stream 0x%04x with stream_type %d", |
| bstream->pid, bstream->stream_type); |
| |
| /* First handle BluRay-specific stream types since there is some overlap |
| * between BluRay and non-BluRay streay type identifiers */ |
| if (program->registration_id == DRF_ID_HDMV) { |
| switch (bstream->stream_type) { |
| case ST_BD_AUDIO_AC3: |
| { |
| const GstMpegtsDescriptor *ac3_desc; |
| |
| /* ATSC ac3 audio descriptor */ |
| ac3_desc = |
| mpegts_get_descriptor_from_stream (bstream, |
| GST_MTS_DESC_AC3_AUDIO_STREAM); |
| if (ac3_desc && DESC_AC_AUDIO_STREAM_bsid (ac3_desc->data) != 16) { |
| GST_LOG ("ac3 audio"); |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-ac3"); |
| } else { |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-eac3"); |
| } |
| break; |
| } |
| case ST_BD_AUDIO_EAC3: |
| case ST_BD_AUDIO_AC3_PLUS: |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-eac3"); |
| break; |
| case ST_BD_AUDIO_AC3_TRUE_HD: |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-true-hd"); |
| stream->target_pes_substream = 0x72; |
| break; |
| case ST_BD_AUDIO_LPCM: |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-private-ts-lpcm"); |
| break; |
| case ST_BD_PGS_SUBPICTURE: |
| is_subpicture = TRUE; |
| caps = gst_caps_new_empty_simple ("subpicture/x-pgs"); |
| sparse = TRUE; |
| break; |
| case ST_BD_AUDIO_DTS_HD: |
| case ST_BD_AUDIO_DTS_HD_MASTER_AUDIO: |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-dts"); |
| stream->target_pes_substream = 0x71; |
| break; |
| } |
| } |
| |
| if (caps) |
| goto done; |
| |
| /* Handle non-BluRay stream types */ |
| switch (bstream->stream_type) { |
| case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG1: |
| case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG2: |
| case ST_PS_VIDEO_MPEG2_DCII: |
| /* FIXME : Use DCII registration code (ETV1 ?) to handle that special |
| * Stream type (ST_PS_VIDEO_MPEG2_DCII) */ |
| /* FIXME : Use video decriptor (0x1) to refine caps with: |
| * * frame_rate |
| * * profile_and_level |
| */ |
| GST_LOG ("mpeg video"); |
| is_video = TRUE; |
| caps = gst_caps_new_simple ("video/mpeg", |
| "mpegversion", G_TYPE_INT, |
| bstream->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG1 ? 1 : 2, |
| "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); |
| |
| break; |
| case GST_MPEGTS_STREAM_TYPE_AUDIO_MPEG1: |
| case GST_MPEGTS_STREAM_TYPE_AUDIO_MPEG2: |
| GST_LOG ("mpeg audio"); |
| is_audio = TRUE; |
| caps = |
| gst_caps_new_simple ("audio/mpeg", "mpegversion", G_TYPE_INT, 1, |
| NULL); |
| /* HDV is always mpeg 1 audio layer 2 */ |
| if (program->registration_id == DRF_ID_TSHV) |
| gst_caps_set_simple (caps, "layer", G_TYPE_INT, 2, NULL); |
| break; |
| case GST_MPEGTS_STREAM_TYPE_PRIVATE_PES_PACKETS: |
| GST_LOG ("private data"); |
| /* FIXME: Move all of this into a common method (there might be other |
| * types also, depending on registratino descriptors also |
| */ |
| desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_AC3); |
| if (desc) { |
| GST_LOG ("ac3 audio"); |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-ac3"); |
| break; |
| } |
| |
| desc = |
| mpegts_get_descriptor_from_stream (bstream, |
| GST_MTS_DESC_DVB_ENHANCED_AC3); |
| if (desc) { |
| GST_LOG ("ac3 audio"); |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-eac3"); |
| break; |
| } |
| desc = |
| mpegts_get_descriptor_from_stream (bstream, |
| GST_MTS_DESC_DVB_TELETEXT); |
| if (desc) { |
| GST_LOG ("teletext"); |
| is_private = TRUE; |
| caps = gst_caps_new_empty_simple ("application/x-teletext"); |
| sparse = TRUE; |
| break; |
| } |
| desc = |
| mpegts_get_descriptor_from_stream (bstream, |
| GST_MTS_DESC_DVB_SUBTITLING); |
| if (desc) { |
| GST_LOG ("subtitling"); |
| is_subpicture = TRUE; |
| caps = gst_caps_new_empty_simple ("subpicture/x-dvb"); |
| sparse = TRUE; |
| break; |
| } |
| |
| switch (bstream->registration_id) { |
| case DRF_ID_DTS1: |
| case DRF_ID_DTS2: |
| case DRF_ID_DTS3: |
| /* SMPTE registered DTS */ |
| is_private = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-dts"); |
| break; |
| case DRF_ID_S302M: |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-smpte-302m"); |
| break; |
| case DRF_ID_OPUS: |
| desc = mpegts_get_descriptor_from_stream (bstream, |
| GST_MTS_DESC_DVB_EXTENSION); |
| if (desc != NULL && desc->tag_extension == 0x80 && desc->length >= 1) { /* User defined (provisional Opus) */ |
| guint8 channel_config_code; |
| GstByteReader br; |
| |
| /* skip tag, length and tag_extension */ |
| gst_byte_reader_init (&br, desc->data + 3, desc->length - 1); |
| channel_config_code = gst_byte_reader_get_uint8_unchecked (&br); |
| |
| if ((channel_config_code & 0x8f) <= 8) { |
| static const guint8 coupled_stream_counts[9] = { |
| 1, 0, 1, 1, 2, 2, 2, 3, 3 |
| }; |
| static const guint8 channel_map_a[8][8] = { |
| {0}, |
| {0, 1}, |
| {0, 2, 1}, |
| {0, 1, 2, 3}, |
| {0, 4, 1, 2, 3}, |
| {0, 4, 1, 2, 3, 5}, |
| {0, 4, 1, 2, 3, 5, 6}, |
| {0, 6, 1, 2, 3, 4, 5, 7}, |
| }; |
| static const guint8 channel_map_b[8][8] = { |
| {0}, |
| {0, 1}, |
| {0, 1, 2}, |
| {0, 1, 2, 3}, |
| {0, 1, 2, 3, 4}, |
| {0, 1, 2, 3, 4, 5}, |
| {0, 1, 2, 3, 4, 5, 6}, |
| {0, 1, 2, 3, 4, 5, 6, 7}, |
| }; |
| |
| gint channels = -1, stream_count, coupled_count, mapping_family; |
| guint8 *channel_mapping = NULL; |
| |
| channels = channel_config_code ? (channel_config_code & 0x0f) : 2; |
| if (channel_config_code == 0 || channel_config_code == 0x80) { |
| /* Dual Mono */ |
| mapping_family = 255; |
| if (channel_config_code == 0) { |
| stream_count = 1; |
| coupled_count = 1; |
| } else { |
| stream_count = 2; |
| coupled_count = 0; |
| } |
| channel_mapping = g_new0 (guint8, channels); |
| memcpy (channel_mapping, &channel_map_a[1], channels); |
| } else if (channel_config_code <= 8) { |
| mapping_family = (channels > 2) ? 1 : 0; |
| stream_count = |
| channel_config_code - |
| coupled_stream_counts[channel_config_code]; |
| coupled_count = coupled_stream_counts[channel_config_code]; |
| if (mapping_family != 0) { |
| channel_mapping = g_new0 (guint8, channels); |
| memcpy (channel_mapping, &channel_map_a[channels - 1], |
| channels); |
| } |
| } else if (channel_config_code >= 0x82 |
| && channel_config_code <= 0x88) { |
| mapping_family = 1; |
| stream_count = channels; |
| coupled_count = 0; |
| channel_mapping = g_new0 (guint8, channels); |
| memcpy (channel_mapping, &channel_map_b[channels - 1], |
| channels); |
| } else if (channel_config_code == 0x81) { |
| if (gst_byte_reader_get_remaining (&br) < 2) { |
| GST_WARNING_OBJECT (demux, |
| "Invalid Opus descriptor with extended channel configuration"); |
| channels = -1; |
| break; |
| } |
| |
| channels = gst_byte_reader_get_uint8_unchecked (&br); |
| mapping_family = gst_byte_reader_get_uint8_unchecked (&br); |
| |
| /* Overwrite values from above */ |
| if (channels == 0) { |
| GST_WARNING_OBJECT (demux, |
| "Invalid Opus descriptor with extended channel configuration"); |
| channels = -1; |
| break; |
| } |
| |
| if (mapping_family == 0 && channels <= 2) { |
| stream_count = channels - coupled_stream_counts[channels]; |
| coupled_count = coupled_stream_counts[channels]; |
| } else { |
| GstBitReader breader; |
| guint8 stream_count_minus_one, coupled_stream_count; |
| gint stream_count_minus_one_len, coupled_stream_count_len; |
| gint channel_mapping_len, i; |
| |
| gst_bit_reader_init (&breader, |
| gst_byte_reader_get_data_unchecked |
| (&br, gst_byte_reader_get_remaining |
| (&br)), gst_byte_reader_get_remaining (&br)); |
| |
| stream_count_minus_one_len = ceil (_gst_log2 (channels)); |
| if (!gst_bit_reader_get_bits_uint8 (&breader, |
| &stream_count_minus_one, |
| stream_count_minus_one_len)) { |
| GST_WARNING_OBJECT (demux, |
| "Invalid Opus descriptor with extended channel configuration"); |
| channels = -1; |
| break; |
| } |
| |
| stream_count = stream_count_minus_one + 1; |
| coupled_stream_count_len = |
| ceil (_gst_log2 (stream_count_minus_one + 2)); |
| |
| if (!gst_bit_reader_get_bits_uint8 (&breader, |
| &coupled_stream_count, coupled_stream_count_len)) { |
| GST_WARNING_OBJECT (demux, |
| "Invalid Opus descriptor with extended channel configuration"); |
| channels = -1; |
| break; |
| } |
| |
| coupled_count = coupled_stream_count; |
| |
| channel_mapping_len = |
| ceil (_gst_log2 (stream_count_minus_one + 1 + |
| coupled_stream_count + 1)); |
| channel_mapping = g_new0 (guint8, channels); |
| for (i = 0; i < channels; i++) { |
| if (!gst_bit_reader_get_bits_uint8 (&breader, |
| &channel_mapping[i], channel_mapping_len)) { |
| GST_WARNING_OBJECT (demux, |
| "Invalid Opus descriptor with extended channel configuration"); |
| break; |
| } |
| } |
| |
| /* error above */ |
| if (i != channels) { |
| channels = -1; |
| g_free (channel_mapping); |
| channel_mapping = NULL; |
| break; |
| } |
| } |
| } else { |
| g_assert_not_reached (); |
| } |
| |
| if (channels != -1) { |
| is_audio = TRUE; |
| caps = |
| gst_codec_utils_opus_create_caps (48000, channels, |
| mapping_family, stream_count, coupled_count, |
| channel_mapping); |
| |
| g_free (channel_mapping); |
| } |
| } else { |
| GST_WARNING_OBJECT (demux, |
| "unexpected channel config code 0x%02x", channel_config_code); |
| } |
| } else { |
| GST_WARNING_OBJECT (demux, "Opus, but no extension descriptor"); |
| } |
| break; |
| case DRF_ID_HEVC: |
| is_video = TRUE; |
| caps = gst_caps_new_simple ("video/x-h265", |
| "stream-format", G_TYPE_STRING, "byte-stream", |
| "alignment", G_TYPE_STRING, "nal", NULL); |
| break; |
| case DRF_ID_KLVA: |
| sparse = TRUE; |
| is_private = TRUE; |
| caps = gst_caps_new_simple ("meta/x-klv", |
| "parsed", G_TYPE_BOOLEAN, TRUE, NULL); |
| break; |
| } |
| if (caps) |
| break; |
| |
| /* hack for itv hd (sid 10510, video pid 3401 */ |
| if (program->program_number == 10510 && bstream->pid == 3401) { |
| is_video = TRUE; |
| caps = gst_caps_new_simple ("video/x-h264", |
| "stream-format", G_TYPE_STRING, "byte-stream", |
| "alignment", G_TYPE_STRING, "nal", NULL); |
| } |
| break; |
| case ST_HDV_AUX_V: |
| /* FIXME : Should only be used with specific PMT registration_descriptor */ |
| /* We don't expose those streams since they're only helper streams */ |
| /* template = gst_static_pad_template_get (&private_template); */ |
| /* name = g_strdup_printf ("private_%04x", bstream->pid); */ |
| /* caps = gst_caps_new_simple ("hdv/aux-v", NULL); */ |
| break; |
| case ST_HDV_AUX_A: |
| /* FIXME : Should only be used with specific PMT registration_descriptor */ |
| /* We don't expose those streams since they're only helper streams */ |
| /* template = gst_static_pad_template_get (&private_template); */ |
| /* name = g_strdup_printf ("private_%04x", bstream->pid); */ |
| /* caps = gst_caps_new_simple ("hdv/aux-a", NULL); */ |
| break; |
| case GST_MPEGTS_STREAM_TYPE_AUDIO_AAC_ADTS: |
| is_audio = TRUE; |
| caps = gst_caps_new_simple ("audio/mpeg", |
| "mpegversion", G_TYPE_INT, 2, |
| "stream-format", G_TYPE_STRING, "adts", NULL); |
| break; |
| case GST_MPEGTS_STREAM_TYPE_AUDIO_AAC_LATM: |
| is_audio = TRUE; |
| caps = gst_caps_new_simple ("audio/mpeg", |
| "mpegversion", G_TYPE_INT, 4, |
| "stream-format", G_TYPE_STRING, "loas", NULL); |
| break; |
| case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG4: |
| is_video = TRUE; |
| caps = gst_caps_new_simple ("video/mpeg", |
| "mpegversion", G_TYPE_INT, 4, |
| "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); |
| break; |
| case GST_MPEGTS_STREAM_TYPE_VIDEO_H264: |
| is_video = TRUE; |
| caps = gst_caps_new_simple ("video/x-h264", |
| "stream-format", G_TYPE_STRING, "byte-stream", |
| "alignment", G_TYPE_STRING, "nal", NULL); |
| break; |
| case GST_MPEGTS_STREAM_TYPE_VIDEO_HEVC: |
| is_video = TRUE; |
| caps = gst_caps_new_simple ("video/x-h265", |
| "stream-format", G_TYPE_STRING, "byte-stream", |
| "alignment", G_TYPE_STRING, "nal", NULL); |
| break; |
| case GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K: |
| is_video = TRUE; |
| desc = |
| mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_J2K_VIDEO); |
| if (desc == NULL) { |
| caps = gst_caps_new_empty_simple ("image/x-jpc"); |
| break; |
| } else { |
| GstByteReader br; |
| guint16 DEN_frame_rate = 0; |
| guint16 NUM_frame_rate = 0; |
| guint8 color_specification = 0; |
| guint8 remaining_8b = 0; |
| gboolean interlaced_video = 0; |
| const gchar *interlace_mode = NULL; |
| const gchar *colorspace = NULL; |
| const gchar *colorimetry_mode = NULL; |
| guint16 profile_and_level G_GNUC_UNUSED; |
| guint32 horizontal_size G_GNUC_UNUSED; |
| guint32 vertical_size G_GNUC_UNUSED; |
| guint32 max_bit_rate G_GNUC_UNUSED; |
| guint32 max_buffer_size G_GNUC_UNUSED; |
| const guint desc_min_length = 24; |
| |
| if (desc->length < desc_min_length) { |
| GST_ERROR |
| ("GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K: descriptor length %d too short", |
| desc->length); |
| return NULL; |
| } |
| |
| /* Skip the descriptor tag and length */ |
| gst_byte_reader_init (&br, desc->data + 2, desc->length); |
| |
| profile_and_level = gst_byte_reader_get_uint16_be_unchecked (&br); |
| horizontal_size = gst_byte_reader_get_uint32_be_unchecked (&br); |
| vertical_size = gst_byte_reader_get_uint32_be_unchecked (&br); |
| max_bit_rate = gst_byte_reader_get_uint32_be_unchecked (&br); |
| max_buffer_size = gst_byte_reader_get_uint32_be_unchecked (&br); |
| DEN_frame_rate = gst_byte_reader_get_uint16_be_unchecked (&br); |
| NUM_frame_rate = gst_byte_reader_get_uint16_be_unchecked (&br); |
| color_specification = gst_byte_reader_get_uint8_unchecked (&br); |
| remaining_8b = gst_byte_reader_get_uint8_unchecked (&br); |
| interlaced_video = remaining_8b & 0x40; |
| /* we don't support demuxing interlaced at the moment */ |
| if (interlaced_video) { |
| GST_ERROR |
| ("GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K: interlaced video not supported"); |
| return NULL; |
| } else { |
| interlace_mode = "progressive"; |
| stream->jp2kInfos.interlace = FALSE; |
| } |
| switch (color_specification) { |
| case GST_MPEGTSDEMUX_JPEG2000_COLORSPEC_SRGB: |
| colorspace = "sRGB"; |
| colorimetry_mode = GST_VIDEO_COLORIMETRY_SRGB; |
| break; |
| case GST_MPEGTSDEMUX_JPEG2000_COLORSPEC_REC601: |
| colorspace = "sYUV"; |
| colorimetry_mode = GST_VIDEO_COLORIMETRY_BT601; |
| break; |
| case GST_MPEGTSDEMUX_JPEG2000_COLORSPEC_REC709: |
| case GST_MPEGTSDEMUX_JPEG2000_COLORSPEC_CIELUV: |
| colorspace = "sYUV"; |
| colorimetry_mode = GST_VIDEO_COLORIMETRY_BT709; |
| break; |
| default: |
| break; |
| } |
| caps = gst_caps_new_simple ("image/x-jpc", |
| "framerate", GST_TYPE_FRACTION, NUM_frame_rate, DEN_frame_rate, |
| "interlace-mode", G_TYPE_STRING, interlace_mode, |
| "colorimetry", G_TYPE_STRING, colorimetry_mode, |
| "colorspace", G_TYPE_STRING, colorspace, NULL); |
| } |
| break; |
| case ST_VIDEO_DIRAC: |
| if (bstream->registration_id == 0x64726163) { |
| GST_LOG ("dirac"); |
| /* dirac in hex */ |
| is_video = TRUE; |
| caps = gst_caps_new_empty_simple ("video/x-dirac"); |
| } |
| break; |
| case ST_PRIVATE_EA: /* Try to detect a VC1 stream */ |
| { |
| gboolean is_vc1 = FALSE; |
| |
| /* Note/FIXME: RP-227 specifies that the registration descriptor |
| * for vc1 can also contain other information, such as profile, |
| * level, alignment, buffer_size, .... */ |
| if (bstream->registration_id == DRF_ID_VC1) |
| is_vc1 = TRUE; |
| if (!is_vc1) { |
| GST_WARNING ("0xea private stream type found but no descriptor " |
| "for VC1. Assuming plain VC1."); |
| } |
| |
| is_video = TRUE; |
| caps = gst_caps_new_simple ("video/x-wmv", |
| "wmvversion", G_TYPE_INT, 3, "format", G_TYPE_STRING, "WVC1", NULL); |
| |
| break; |
| } |
| case ST_PS_AUDIO_AC3: |
| /* DVB_ENHANCED_AC3 */ |
| desc = |
| mpegts_get_descriptor_from_stream (bstream, |
| GST_MTS_DESC_DVB_ENHANCED_AC3); |
| if (desc) { |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-eac3"); |
| break; |
| } |
| |
| /* If stream has ac3 descriptor |
| * OR program is ATSC (GA94) |
| * OR stream registration is AC-3 |
| * then it's regular AC3 */ |
| if (bstream->registration_id == DRF_ID_AC3 || |
| program->registration_id == DRF_ID_GA94 || |
| mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_AC3)) { |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-ac3"); |
| break; |
| } |
| |
| GST_WARNING ("AC3 stream type found but no guaranteed " |
| "way found to differentiate between AC3 and EAC3. " |
| "Assuming plain AC3."); |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-ac3"); |
| break; |
| case ST_PS_AUDIO_EAC3: |
| { |
| /* ATSC_ENHANCED_AC3 */ |
| if (bstream->registration_id == DRF_ID_EAC3 || |
| mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_ATSC_EAC3)) { |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-eac3"); |
| break; |
| } |
| |
| GST_ELEMENT_WARNING (demux, STREAM, DEMUX, |
| ("Assuming ATSC E-AC3 audio stream."), |
| ("ATSC E-AC3 stream type found but no guarantee way found to " |
| "differentiate among other standards (DVB, ISDB and etc..)")); |
| |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-eac3"); |
| break; |
| } |
| case ST_PS_AUDIO_LPCM2: |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-private2-lpcm"); |
| break; |
| case ST_PS_AUDIO_DTS: |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-dts"); |
| break; |
| case ST_PS_AUDIO_LPCM: |
| is_audio = TRUE; |
| caps = gst_caps_new_empty_simple ("audio/x-lpcm"); |
| break; |
| case ST_PS_DVD_SUBPICTURE: |
| is_subpicture = TRUE; |
| caps = gst_caps_new_empty_simple ("subpicture/x-dvd"); |
| sparse = TRUE; |
| break; |
| case 0x42: |
| /* hack for Chinese AVS video stream which use 0x42 as stream_id |
| * NOTE: this is unofficial and within the ISO reserved range. */ |
| is_video = TRUE; |
| caps = gst_caps_new_empty_simple ("video/x-cavs"); |
| break; |
| default: |
| GST_DEBUG ("Non-media stream (stream_type:0x%x). Not creating pad", |
| bstream->stream_type); |
| break; |
| } |
| |
| done: |
| if (caps) { |
| if (is_audio) { |
| template = gst_static_pad_template_get (&audio_template); |
| name = |
| g_strdup_printf ("audio_%01x_%04x", demux->program_generation, |
| bstream->pid); |
| gst_stream_set_stream_type (bstream->stream_object, |
| GST_STREAM_TYPE_AUDIO); |
| } else if (is_video) { |
| template = gst_static_pad_template_get (&video_template); |
| name = |
| g_strdup_printf ("video_%01x_%04x", demux->program_generation, |
| bstream->pid); |
| gst_stream_set_stream_type (bstream->stream_object, |
| GST_STREAM_TYPE_VIDEO); |
| } else if (is_private) { |
| template = gst_static_pad_template_get (&private_template); |
| name = |
| g_strdup_printf ("private_%01x_%04x", demux->program_generation, |
| bstream->pid); |
| } else if (is_subpicture) { |
| template = gst_static_pad_template_get (&subpicture_template); |
| name = |
| g_strdup_printf ("subpicture_%01x_%04x", demux->program_generation, |
| bstream->pid); |
| gst_stream_set_stream_type (bstream->stream_object, GST_STREAM_TYPE_TEXT); |
| } else |
| g_assert_not_reached (); |
| |
| } |
| |
| if (template && name && caps) { |
| GstEvent *event; |
| const gchar *stream_id; |
| |
| GST_LOG ("stream:%p creating pad with name %s and caps %" GST_PTR_FORMAT, |
| stream, name, caps); |
| pad = gst_pad_new_from_template (template, name); |
| gst_pad_set_active (pad, TRUE); |
| gst_pad_use_fixed_caps (pad); |
| stream_id = gst_stream_get_stream_id (bstream->stream_object); |
| |
| event = gst_pad_get_sticky_event (base->sinkpad, GST_EVENT_STREAM_START, 0); |
| if (event) { |
| if (gst_event_parse_group_id (event, &demux->group_id)) |
| demux->have_group_id = TRUE; |
| else |
| demux->have_group_id = FALSE; |
| gst_event_unref (event); |
| } else if (!demux->have_group_id) { |
| demux->have_group_id = TRUE; |
| demux->group_id = gst_util_group_id_next (); |
| } |
| event = gst_event_new_stream_start (stream_id); |
| gst_event_set_stream (event, bstream->stream_object); |
| if (demux->have_group_id) |
| gst_event_set_group_id (event, demux->group_id); |
| if (sparse) { |
| gst_event_set_stream_flags (event, GST_STREAM_FLAG_SPARSE); |
| gst_stream_set_stream_flags (bstream->stream_object, |
| GST_STREAM_FLAG_SPARSE); |
| } |
| stream->sparse = sparse; |
| gst_stream_set_caps (bstream->stream_object, caps); |
| if (!stream->taglist) |
| stream->taglist = gst_tag_list_new_empty (); |
| gst_pb_utils_add_codec_description_to_tag_list (stream->taglist, NULL, |
| caps); |
| gst_stream_set_tags (bstream->stream_object, stream->taglist); |
| |
| gst_pad_push_event (pad, event); |
| gst_pad_set_caps (pad, caps); |
| gst_pad_set_query_function (pad, gst_ts_demux_srcpad_query); |
| gst_pad_set_event_function (pad, gst_ts_demux_srcpad_event); |
| } |
| |
| g_free (name); |
| if (template) |
| gst_object_unref (template); |
| if (caps) |
| gst_caps_unref (caps); |
| |
| return pad; |
| } |
| |
| static gboolean |
| gst_ts_demux_stream_added (MpegTSBase * base, MpegTSBaseStream * bstream, |
| MpegTSBaseProgram * program) |
| { |
| GstTSDemux *demux = (GstTSDemux *) base; |
| TSDemuxStream *stream = (TSDemuxStream *) bstream; |
| |
| if (!stream->pad) { |
| /* Create the pad */ |
| if (bstream->stream_type != 0xff) { |
| stream->pad = create_pad_for_stream (base, bstream, program); |
| if (stream->pad) |
| gst_flow_combiner_add_pad (demux->flowcombiner, stream->pad); |
| } |
| |
| if (base->mode != BASE_MODE_PUSHING |
| && bstream->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_H264) { |
| stream->scan_function = |
| (GstTsDemuxKeyFrameScanFunction) scan_keyframe_h264; |
| } else { |
| stream->scan_function = NULL; |
| } |
| |
| stream->active = FALSE; |
| |
| stream->need_newsegment = TRUE; |
| /* Reset segment if we're not doing an accurate seek */ |
| demux->reset_segment = (!(demux->segment.flags & GST_SEEK_FLAG_ACCURATE)); |
| stream->needs_keyframe = FALSE; |
| stream->discont = TRUE; |
| stream->pts = GST_CLOCK_TIME_NONE; |
| stream->dts = GST_CLOCK_TIME_NONE; |
| stream->first_pts = GST_CLOCK_TIME_NONE; |
| stream->raw_pts = -1; |
| stream->raw_dts = -1; |
| stream->pending_ts = TRUE; |
| stream->nb_out_buffers = 0; |
| stream->gap_ref_buffers = 0; |
| stream->gap_ref_pts = GST_CLOCK_TIME_NONE; |
| /* Only wait for a valid timestamp if we have a PCR_PID */ |
| stream->pending_ts = program->pcr_pid < 0x1fff; |
| stream->continuity_counter = CONTINUITY_UNSET; |
| } |
| |
| return (stream->pad != NULL); |
| } |
| |
| static void |
| tsdemux_h264_parsing_info_clear (TSDemuxH264ParsingInfos * h264infos) |
| { |
| clear_simple_buffer (&h264infos->framedata); |
| |
| if (h264infos->parser) { |
| gst_h264_nal_parser_free (h264infos->parser); |
| gst_byte_writer_free (h264infos->sps); |
| gst_byte_writer_free (h264infos->pps); |
| gst_byte_writer_free (h264infos->sei); |
| } |
| } |
| |
| static void |
| gst_ts_demux_stream_removed (MpegTSBase * base, MpegTSBaseStream * bstream) |
| { |
| TSDemuxStream *stream = (TSDemuxStream *) bstream; |
| |
| if (stream->pad) { |
| gst_flow_combiner_remove_pad (GST_TS_DEMUX_CAST (base)->flowcombiner, |
| stream->pad); |
| if (stream->active) { |
| |
| if (gst_pad_is_active (stream->pad)) { |
| /* Flush out all data */ |
| GST_DEBUG_OBJECT (stream->pad, "Flushing out pending data"); |
| gst_ts_demux_push_pending_data ((GstTSDemux *) base, stream, NULL); |
| |
| GST_DEBUG_OBJECT (stream->pad, "Pushing out EOS"); |
| gst_pad_push_event (stream->pad, gst_event_new_eos ()); |
| gst_pad_set_active (stream->pad, FALSE); |
| } |
| |
| GST_DEBUG_OBJECT (stream->pad, "Removing pad"); |
| gst_element_remove_pad (GST_ELEMENT_CAST (base), stream->pad); |
| stream->active = FALSE; |
| } else { |
| gst_object_unref (stream->pad); |
| } |
| stream->pad = NULL; |
| } |
| |
| gst_ts_demux_stream_flush (stream, GST_TS_DEMUX_CAST (base), TRUE); |
| |
| if (stream->taglist != NULL) { |
| gst_tag_list_unref (stream->taglist); |
| stream->taglist = NULL; |
| } |
| |
| tsdemux_h264_parsing_info_clear (&stream->h264infos); |
| } |
| |
| static void |
| activate_pad_for_stream (GstTSDemux * tsdemux, TSDemuxStream * stream) |
| { |
| if (stream->pad) { |
| GST_DEBUG_OBJECT (tsdemux, "Activating pad %s:%s for stream %p", |
| GST_DEBUG_PAD_NAME (stream->pad), stream); |
| gst_element_add_pad ((GstElement *) tsdemux, stream->pad); |
| stream->active = TRUE; |
| GST_DEBUG_OBJECT (stream->pad, "done adding pad"); |
| } else if (((MpegTSBaseStream *) stream)->stream_type != 0xff) { |
| GST_DEBUG_OBJECT (tsdemux, |
| "stream %p (pid 0x%04x, type:0x%02x) has no pad", stream, |
| ((MpegTSBaseStream *) stream)->pid, |
| ((MpegTSBaseStream *) stream)->stream_type); |
| } |
| } |
| |
| static void |
| gst_ts_demux_stream_flush (TSDemuxStream * stream, GstTSDemux * tsdemux, |
| gboolean hard) |
| { |
| GST_DEBUG ("flushing stream %p", stream); |
| |
| g_free (stream->data); |
| stream->data = NULL; |
| stream->state = PENDING_PACKET_EMPTY; |
| stream->expected_size = 0; |
| stream->allocated_size = 0; |
| stream->current_size = 0; |
| stream->discont = TRUE; |
| stream->pts = GST_CLOCK_TIME_NONE; |
| stream->dts = GST_CLOCK_TIME_NONE; |
| stream->raw_pts = -1; |
| stream->raw_dts = -1; |
| stream->pending_ts = TRUE; |
| stream->nb_out_buffers = 0; |
| stream->gap_ref_buffers = 0; |
| stream->gap_ref_pts = GST_CLOCK_TIME_NONE; |
| stream->continuity_counter = CONTINUITY_UNSET; |
| |
| if (G_UNLIKELY (stream->pending)) { |
| GList *tmp; |
| |
| GST_DEBUG ("clearing pending %p", stream); |
| for (tmp = stream->pending; tmp; tmp = tmp->next) { |
| PendingBuffer *pend = (PendingBuffer *) tmp->data; |
| gst_buffer_unref (pend->buffer); |
| g_slice_free (PendingBuffer, pend); |
| } |
| g_list_free (stream->pending); |
| stream->pending = NULL; |
| } |
| |
| if (hard) { |
| stream->first_pts = GST_CLOCK_TIME_NONE; |
| stream->need_newsegment = TRUE; |
| } |
| } |
| |
| static void |
| gst_ts_demux_flush_streams (GstTSDemux * demux, gboolean hard) |
| { |
| GList *walk; |
| if (!demux->program) |
| return; |
| |
| for (walk = demux->program->stream_list; walk; walk = g_list_next (walk)) |
| gst_ts_demux_stream_flush (walk->data, demux, hard); |
| } |
| |
| static gboolean |
| gst_ts_demux_can_remove_program (MpegTSBase * base, MpegTSBaseProgram * program) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX (base); |
| |
| /* If it's our current active program, we return FALSE, we'll deactivate it |
| * ourselves when the next program gets activated */ |
| if (demux->program == program) { |
| GST_DEBUG |
| ("Attempting to remove current program, delaying until new program gets activated"); |
| demux->previous_program = program; |
| demux->program_number = -1; |
| return FALSE; |
| } |
| return TRUE; |
| } |
| |
| static void |
| gst_ts_demux_update_program (MpegTSBase * base, MpegTSBaseProgram * program) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX (base); |
| GList *tmp; |
| |
| GST_DEBUG ("Updating program %d", program->program_number); |
| /* Emit collection message */ |
| gst_element_post_message ((GstElement *) base, |
| gst_message_new_stream_collection ((GstObject *) base, |
| program->collection)); |
| |
| /* Add all streams, then fire no-more-pads */ |
| for (tmp = program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = (TSDemuxStream *) tmp->data; |
| if (!stream->pad) { |
| activate_pad_for_stream (demux, stream); |
| if (stream->sparse) { |
| /* force sending of pending sticky events which have been stored on the |
| * pad already and which otherwise would only be sent on the first buffer |
| * or serialized event (which means very late in case of subtitle streams), |
| * and playsink waits for stream-start or another serialized event */ |
| GST_DEBUG_OBJECT (stream->pad, "sparse stream, pushing GAP event"); |
| gst_pad_push_event (stream->pad, gst_event_new_gap (0, 0)); |
| } |
| } |
| } |
| } |
| |
| static void |
| gst_ts_demux_program_started (MpegTSBase * base, MpegTSBaseProgram * program) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX (base); |
| |
| GST_DEBUG ("Current program %d, new program %d requested program %d", |
| (gint) demux->program_number, program->program_number, |
| demux->requested_program_number); |
| |
| if (demux->requested_program_number == program->program_number || |
| (demux->requested_program_number == -1 && demux->program_number == -1)) { |
| GList *tmp; |
| gboolean have_pads = FALSE; |
| |
| GST_LOG ("program %d started", program->program_number); |
| demux->program_number = program->program_number; |
| demux->program = program; |
| |
| /* Increment the program_generation counter */ |
| demux->program_generation = (demux->program_generation + 1) & 0xf; |
| |
| /* Emit collection message */ |
| gst_element_post_message ((GstElement *) base, |
| gst_message_new_stream_collection ((GstObject *) base, |
| program->collection)); |
| |
| /* If this is not the initial program, we need to calculate |
| * a new segment */ |
| if (demux->segment_event) { |
| gst_event_unref (demux->segment_event); |
| demux->segment_event = NULL; |
| } |
| |
| /* DRAIN ALL STREAMS FIRST ! */ |
| if (demux->previous_program) { |
| GList *tmp; |
| GST_DEBUG_OBJECT (demux, "Draining previous program"); |
| for (tmp = demux->previous_program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = (TSDemuxStream *) tmp->data; |
| if (stream->pad) |
| gst_ts_demux_push_pending_data (demux, stream, |
| demux->previous_program); |
| } |
| } |
| |
| /* Add all streams, then fire no-more-pads */ |
| for (tmp = program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = (TSDemuxStream *) tmp->data; |
| activate_pad_for_stream (demux, stream); |
| if (stream->pad) |
| have_pads = TRUE; |
| } |
| |
| /* If there was a previous program, now is the time to deactivate it |
| * and remove old pads (including pushing EOS) */ |
| if (demux->previous_program) { |
| GST_DEBUG ("Deactivating previous program"); |
| mpegts_base_deactivate_and_free_program (base, demux->previous_program); |
| demux->previous_program = NULL; |
| } |
| |
| if (!have_pads) { |
| /* If we had no pads, this stream is likely corrupted or unsupported and |
| * there's not much we can do at this point */ |
| GST_ELEMENT_ERROR (demux, STREAM, WRONG_TYPE, |
| ("This stream contains no valid or supported streams."), |
| ("activating program but got no pads")); |
| return; |
| } |
| |
| /* If any of the stream is sparse, push a GAP event before anything else |
| * This is done here, and not in activate_pad_for_stream() because pushing |
| * a GAP event *is* considering data, and we want to ensure the (potential) |
| * old pads are all removed before we push any data on the new ones */ |
| for (tmp = program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = (TSDemuxStream *) tmp->data; |
| if (stream->sparse) { |
| /* force sending of pending sticky events which have been stored on the |
| * pad already and which otherwise would only be sent on the first buffer |
| * or serialized event (which means very late in case of subtitle streams), |
| * and playsink waits for stream-start or another serialized event */ |
| GST_DEBUG_OBJECT (stream->pad, "sparse stream, pushing GAP event"); |
| gst_pad_push_event (stream->pad, gst_event_new_gap (0, 0)); |
| } |
| } |
| |
| gst_element_no_more_pads ((GstElement *) demux); |
| } |
| } |
| |
| static void |
| gst_ts_demux_program_stopped (MpegTSBase * base, MpegTSBaseProgram * program) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX (base); |
| |
| if (demux->program == program) { |
| demux->program = NULL; |
| demux->program_number = -1; |
| } |
| } |
| |
| |
| static inline void |
| gst_ts_demux_record_pts (GstTSDemux * demux, TSDemuxStream * stream, |
| guint64 pts, guint64 offset) |
| { |
| MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; |
| |
| stream->raw_pts = pts; |
| if (pts == -1) { |
| stream->pts = GST_CLOCK_TIME_NONE; |
| return; |
| } |
| |
| GST_LOG ("pid 0x%04x raw pts:%" G_GUINT64_FORMAT " at offset %" |
| G_GUINT64_FORMAT, bs->pid, pts, offset); |
| |
| /* Compute PTS in GstClockTime */ |
| stream->pts = |
| mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), |
| MPEGTIME_TO_GSTTIME (pts), demux->program->pcr_pid); |
| |
| GST_LOG ("pid 0x%04x Stored PTS %" G_GUINT64_FORMAT, bs->pid, stream->pts); |
| |
| if (G_UNLIKELY (demux->emit_statistics)) { |
| GstStructure *st; |
| st = gst_structure_new_id_empty (QUARK_TSDEMUX); |
| gst_structure_id_set (st, |
| QUARK_PID, G_TYPE_UINT, bs->pid, |
| QUARK_OFFSET, G_TYPE_UINT64, offset, QUARK_PTS, G_TYPE_UINT64, pts, |
| NULL); |
| gst_element_post_message (GST_ELEMENT_CAST (demux), |
| gst_message_new_element (GST_OBJECT (demux), st)); |
| } |
| } |
| |
| static inline void |
| gst_ts_demux_record_dts (GstTSDemux * demux, TSDemuxStream * stream, |
| guint64 dts, guint64 offset) |
| { |
| MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; |
| |
| stream->raw_dts = dts; |
| if (dts == -1) { |
| stream->dts = GST_CLOCK_TIME_NONE; |
| return; |
| } |
| |
| GST_LOG ("pid 0x%04x raw dts:%" G_GUINT64_FORMAT " at offset %" |
| G_GUINT64_FORMAT, bs->pid, dts, offset); |
| |
| /* Compute DTS in GstClockTime */ |
| stream->dts = |
| mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), |
| MPEGTIME_TO_GSTTIME (dts), demux->program->pcr_pid); |
| |
| GST_LOG ("pid 0x%04x Stored DTS %" G_GUINT64_FORMAT, bs->pid, stream->dts); |
| |
| if (G_UNLIKELY (demux->emit_statistics)) { |
| GstStructure *st; |
| st = gst_structure_new_id_empty (QUARK_TSDEMUX); |
| gst_structure_id_set (st, |
| QUARK_PID, G_TYPE_UINT, bs->pid, |
| QUARK_OFFSET, G_TYPE_UINT64, offset, QUARK_DTS, G_TYPE_UINT64, dts, |
| NULL); |
| gst_element_post_message (GST_ELEMENT_CAST (demux), |
| gst_message_new_element (GST_OBJECT (demux), st)); |
| } |
| } |
| |
| /* This is called when we haven't got a valid initial PTS/DTS on all streams */ |
| static gboolean |
| check_pending_buffers (GstTSDemux * demux) |
| { |
| gboolean have_observation = FALSE; |
| /* The biggest offset */ |
| guint64 offset = 0; |
| GList *tmp; |
| gboolean have_only_sparse = TRUE; |
| |
| /* 0. Do we only have sparse stream */ |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *tmpstream = (TSDemuxStream *) tmp->data; |
| |
| if (!tmpstream->sparse) { |
| have_only_sparse = FALSE; |
| break; |
| } |
| } |
| |
| /* 1. Go over all streams */ |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *tmpstream = (TSDemuxStream *) tmp->data; |
| /* 1.1 check if at least one stream got a valid DTS */ |
| if (have_only_sparse || !tmpstream->sparse) { |
| if ((tmpstream->raw_dts != -1 && tmpstream->dts != GST_CLOCK_TIME_NONE) || |
| (tmpstream->raw_pts != -1 && tmpstream->pts != GST_CLOCK_TIME_NONE)) { |
| have_observation = TRUE; |
| break; |
| } |
| } |
| } |
| |
| /* 2. If we don't have a valid value yet, break out */ |
| if (have_observation == FALSE) |
| return FALSE; |
| |
| /* 3. Go over all streams that have current/pending data */ |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *tmpstream = (TSDemuxStream *) tmp->data; |
| PendingBuffer *pend; |
| guint64 firstval, lastval, ts; |
| |
| /* 3.1 Calculate the offset between current DTS and first DTS */ |
| if (tmpstream->pending == NULL || tmpstream->state == PENDING_PACKET_EMPTY) |
| continue; |
| /* If we don't have any pending data, the offset is 0 for this stream */ |
| if (tmpstream->pending == NULL) |
| break; |
| if (tmpstream->raw_dts != -1) |
| lastval = tmpstream->raw_dts; |
| else if (tmpstream->raw_pts != -1) |
| lastval = tmpstream->raw_pts; |
| else { |
| GST_WARNING ("Don't have a last DTS/PTS to use for offset recalculation"); |
| continue; |
| } |
| pend = tmpstream->pending->data; |
| if (pend->dts != -1) |
| firstval = pend->dts; |
| else if (pend->pts != -1) |
| firstval = pend->pts; |
| else { |
| GST_WARNING |
| ("Don't have a first DTS/PTS to use for offset recalculation"); |
| continue; |
| } |
| /* 3.2 Add to the offset the report TS for the current DTS */ |
| ts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), |
| MPEGTIME_TO_GSTTIME (lastval), demux->program->pcr_pid); |
| if (ts == GST_CLOCK_TIME_NONE) { |
| GST_WARNING ("THIS SHOULD NOT HAPPEN !"); |
| continue; |
| } |
| ts += MPEGTIME_TO_GSTTIME (lastval - firstval); |
| /* 3.3 If that offset is bigger than the current offset, store it */ |
| if (ts > offset) |
| offset = ts; |
| } |
| |
| GST_DEBUG ("New initial pcr_offset %" GST_TIME_FORMAT, |
| GST_TIME_ARGS (offset)); |
| |
| /* 4. Set the offset on the packetizer */ |
| mpegts_packetizer_set_current_pcr_offset (MPEG_TS_BASE_PACKETIZER (demux), |
| offset, demux->program->pcr_pid); |
| |
| /* 4. Go over all streams */ |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = (TSDemuxStream *) tmp->data; |
| |
| stream->pending_ts = FALSE; |
| /* 4.1 Set pending_ts for FALSE */ |
| |
| /* 4.2 Recalculate PTS/DTS (in running time) for pending data */ |
| if (stream->pending) { |
| GList *tmp2; |
| for (tmp2 = stream->pending; tmp2; tmp2 = tmp2->next) { |
| PendingBuffer *pend = (PendingBuffer *) tmp2->data; |
| if (pend->pts != -1) |
| GST_BUFFER_PTS (pend->buffer) = |
| mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), |
| MPEGTIME_TO_GSTTIME (pend->pts), demux->program->pcr_pid); |
| if (pend->dts != -1) |
| GST_BUFFER_DTS (pend->buffer) = |
| mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), |
| MPEGTIME_TO_GSTTIME (pend->dts), demux->program->pcr_pid); |
| /* 4.2.2 Set first_pts to TS of lowest PTS (for segment) */ |
| if (stream->first_pts == GST_CLOCK_TIME_NONE) { |
| if (GST_BUFFER_PTS (pend->buffer) != GST_CLOCK_TIME_NONE) |
| stream->first_pts = GST_BUFFER_PTS (pend->buffer); |
| else if (GST_BUFFER_DTS (pend->buffer) != GST_CLOCK_TIME_NONE) |
| stream->first_pts = GST_BUFFER_DTS (pend->buffer); |
| } |
| } |
| } |
| /* Recalculate PTS/DTS (in running time) for current data */ |
| if (stream->state != PENDING_PACKET_EMPTY) { |
| if (stream->raw_pts != -1) { |
| stream->pts = |
| mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), |
| MPEGTIME_TO_GSTTIME (stream->raw_pts), demux->program->pcr_pid); |
| if (stream->first_pts == GST_CLOCK_TIME_NONE) |
| stream->first_pts = stream->pts; |
| } |
| if (stream->raw_dts != -1) { |
| stream->dts = |
| mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), |
| MPEGTIME_TO_GSTTIME (stream->raw_dts), demux->program->pcr_pid); |
| if (stream->first_pts == GST_CLOCK_TIME_NONE) |
| stream->first_pts = stream->dts; |
| } |
| } |
| } |
| |
| return TRUE; |
| } |
| |
| static void |
| gst_ts_demux_parse_pes_header (GstTSDemux * demux, TSDemuxStream * stream, |
| guint8 * data, guint32 length, guint64 bufferoffset) |
| { |
| PESHeader header; |
| PESParsingResult parseres; |
| |
| GST_MEMDUMP ("Header buffer", data, MIN (length, 32)); |
| |
| parseres = mpegts_parse_pes_header (data, length, &header); |
| if (G_UNLIKELY (parseres == PES_PARSING_NEED_MORE)) |
| goto discont; |
| if (G_UNLIKELY (parseres == PES_PARSING_BAD)) { |
| GST_WARNING ("Error parsing PES header. pid: 0x%x stream_type: 0x%x", |
| stream->stream.pid, stream->stream.stream_type); |
| goto discont; |
| } |
| |
| if (stream->target_pes_substream != 0 |
| && header.stream_id_extension != stream->target_pes_substream) { |
| GST_DEBUG ("Skipping unwanted substream"); |
| goto discont; |
| } |
| |
| gst_ts_demux_record_dts (demux, stream, header.DTS, bufferoffset); |
| gst_ts_demux_record_pts (demux, stream, header.PTS, bufferoffset); |
| if (G_UNLIKELY (stream->pending_ts && |
| (stream->pts != GST_CLOCK_TIME_NONE |
| || stream->dts != GST_CLOCK_TIME_NONE))) { |
| GST_DEBUG ("Got pts/dts update, rechecking all streams"); |
| check_pending_buffers (demux); |
| } else if (stream->first_pts == GST_CLOCK_TIME_NONE) { |
| if (GST_CLOCK_TIME_IS_VALID (stream->pts)) |
| stream->first_pts = stream->pts; |
| else if (GST_CLOCK_TIME_IS_VALID (stream->dts)) |
| stream->first_pts = stream->dts; |
| } |
| |
| GST_DEBUG_OBJECT (demux, |
| "stream PTS %" GST_TIME_FORMAT " DTS %" GST_TIME_FORMAT, |
| GST_TIME_ARGS (stream->pts), GST_TIME_ARGS (stream->dts)); |
| |
| /* Remove PES headers */ |
| GST_DEBUG ("Moving data forward by %d bytes (packet_size:%d, have:%d)", |
| header.header_size, header.packet_length, length); |
| stream->expected_size = header.packet_length; |
| if (stream->expected_size) { |
| if (G_LIKELY (stream->expected_size > header.header_size)) { |
| stream->expected_size -= header.header_size; |
| } else { |
| /* next packet will have to complete this one */ |
| GST_WARNING ("invalid header and packet size combination, empty packet"); |
| stream->expected_size = 0; |
| } |
| } |
| data += header.header_size; |
| length -= header.header_size; |
| |
| /* Create the output buffer */ |
| if (stream->expected_size) |
| stream->allocated_size = MAX (stream->expected_size, length); |
| else |
| stream->allocated_size = MAX (8192, length); |
| |
| g_assert (stream->data == NULL); |
| stream->data = g_malloc (stream->allocated_size); |
| memcpy (stream->data, data, length); |
| stream->current_size = length; |
| |
| stream->state = PENDING_PACKET_BUFFER; |
| |
| return; |
| |
| discont: |
| stream->state = PENDING_PACKET_DISCONT; |
| return; |
| } |
| |
| /* ONLY CALL THIS: |
| * * WITH packet->payload != NULL |
| * * WITH pending/current flushed out if beginning of new PES packet |
| */ |
| static inline void |
| gst_ts_demux_queue_data (GstTSDemux * demux, TSDemuxStream * stream, |
| MpegTSPacketizerPacket * packet) |
| { |
| guint8 *data; |
| guint size; |
| guint8 cc = FLAGS_CONTINUITY_COUNTER (packet->scram_afc_cc); |
| |
| GST_LOG ("pid: 0x%04x state:%d", stream->stream.pid, stream->state); |
| |
| size = packet->data_end - packet->payload; |
| data = packet->payload; |
| |
| if (stream->continuity_counter == CONTINUITY_UNSET) { |
| GST_DEBUG ("CONTINUITY: Initialize to %d", cc); |
| } else if ((cc == stream->continuity_counter + 1 || |
| (stream->continuity_counter == MAX_CONTINUITY && cc == 0))) { |
| GST_LOG ("CONTINUITY: Got expected %d", cc); |
| } else { |
| GST_WARNING ("CONTINUITY: Mismatch packet %d, stream %d", |
| cc, stream->continuity_counter); |
| if (stream->state != PENDING_PACKET_EMPTY) |
| stream->state = PENDING_PACKET_DISCONT; |
| } |
| stream->continuity_counter = cc; |
| |
| if (stream->state == PENDING_PACKET_EMPTY) { |
| if (G_UNLIKELY (!packet->payload_unit_start_indicator)) { |
| stream->state = PENDING_PACKET_DISCONT; |
| GST_DEBUG ("Didn't get the first packet of this PES"); |
| } else { |
| GST_LOG ("EMPTY=>HEADER"); |
| stream->state = PENDING_PACKET_HEADER; |
| } |
| } |
| |
| switch (stream->state) { |
| case PENDING_PACKET_HEADER: |
| { |
| GST_LOG ("HEADER: Parsing PES header"); |
| |
| /* parse the header */ |
| gst_ts_demux_parse_pes_header (demux, stream, data, size, packet->offset); |
| break; |
| } |
| case PENDING_PACKET_BUFFER: |
| { |
| GST_LOG ("BUFFER: appending data"); |
| if (G_UNLIKELY (stream->current_size + size > stream->allocated_size)) { |
| GST_LOG ("resizing buffer"); |
| do { |
| stream->allocated_size *= 2; |
| } while (stream->current_size + size > stream->allocated_size); |
| stream->data = g_realloc (stream->data, stream->allocated_size); |
| } |
| memcpy (stream->data + stream->current_size, data, size); |
| stream->current_size += size; |
| break; |
| } |
| case PENDING_PACKET_DISCONT: |
| { |
| GST_LOG ("DISCONT: not storing/pushing"); |
| if (G_UNLIKELY (stream->data)) { |
| g_free (stream->data); |
| stream->data = NULL; |
| } |
| stream->continuity_counter = CONTINUITY_UNSET; |
| break; |
| } |
| default: |
| break; |
| } |
| |
| return; |
| } |
| |
| static void |
| calculate_and_push_newsegment (GstTSDemux * demux, TSDemuxStream * stream, |
| MpegTSBaseProgram * target_program) |
| { |
| MpegTSBase *base = (MpegTSBase *) demux; |
| GstClockTime lowest_pts = GST_CLOCK_TIME_NONE; |
| GstClockTime firstts = 0; |
| GList *tmp; |
| |
| GST_DEBUG ("Creating new newsegment for stream %p", stream); |
| |
| if (target_program == NULL) |
| target_program = demux->program; |
| |
| /* Speedup : if we don't need to calculate anything, go straight to pushing */ |
| if (demux->segment_event) |
| goto push_new_segment; |
| |
| /* Calculate the 'new_start' value, used for newsegment */ |
| for (tmp = target_program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *pstream = (TSDemuxStream *) tmp->data; |
| |
| if (GST_CLOCK_TIME_IS_VALID (pstream->first_pts)) { |
| if (!GST_CLOCK_TIME_IS_VALID (lowest_pts) |
| || pstream->first_pts < lowest_pts) |
| lowest_pts = pstream->first_pts; |
| } |
| } |
| if (GST_CLOCK_TIME_IS_VALID (lowest_pts)) |
| firstts = lowest_pts; |
| GST_DEBUG ("lowest_pts %" G_GUINT64_FORMAT " => clocktime %" GST_TIME_FORMAT, |
| lowest_pts, GST_TIME_ARGS (firstts)); |
| |
| if (demux->segment.format != GST_FORMAT_TIME || demux->reset_segment) { |
| /* It will happen only if it's first program or after flushes. */ |
| GST_DEBUG ("Calculating actual segment"); |
| if (base->segment.format == GST_FORMAT_TIME) { |
| /* Try to recover segment info from base if it's in TIME format */ |
| demux->segment = base->segment; |
| } else { |
| /* Start from the first ts/pts */ |
| GstClockTime base = |
| demux->segment.base + demux->segment.position - demux->segment.start; |
| gst_segment_init (&demux->segment, GST_FORMAT_TIME); |
| demux->segment.start = firstts; |
| demux->segment.stop = GST_CLOCK_TIME_NONE; |
| demux->segment.position = firstts; |
| demux->segment.time = firstts; |
| demux->segment.rate = demux->rate; |
| demux->segment.base = base; |
| } |
| } else if (demux->segment.start < firstts) { |
| /* Take into account the offset to the first buffer timestamp */ |
| if (demux->segment.rate > 0) { |
| demux->segment.start = firstts; |
| |
| if (GST_CLOCK_TIME_IS_VALID (demux->segment.stop)) |
| demux->segment.stop += firstts - demux->segment.start; |
| demux->segment.position = firstts; |
| } |
| } |
| |
| if (!demux->segment_event) { |
| demux->segment_event = gst_event_new_segment (&demux->segment); |
| GST_EVENT_SEQNUM (demux->segment_event) = base->last_seek_seqnum; |
| } |
| |
| push_new_segment: |
| for (tmp = target_program->stream_list; tmp; tmp = tmp->next) { |
| stream = (TSDemuxStream *) tmp->data; |
| if (stream->pad == NULL) |
| continue; |
| |
| if (demux->segment_event) { |
| GST_DEBUG_OBJECT (stream->pad, "Pushing newsegment event"); |
| gst_event_ref (demux->segment_event); |
| gst_pad_push_event (stream->pad, demux->segment_event); |
| } |
| |
| if (demux->global_tags) { |
| gst_pad_push_event (stream->pad, |
| gst_event_new_tag (gst_tag_list_ref (demux->global_tags))); |
| } |
| |
| /* Push pending tags */ |
| if (stream->taglist) { |
| GST_DEBUG_OBJECT (stream->pad, "Sending tags %" GST_PTR_FORMAT, |
| stream->taglist); |
| gst_pad_push_event (stream->pad, gst_event_new_tag (stream->taglist)); |
| stream->taglist = NULL; |
| } |
| |
| stream->need_newsegment = FALSE; |
| } |
| } |
| |
| static void |
| gst_ts_demux_check_and_sync_streams (GstTSDemux * demux, GstClockTime time) |
| { |
| GList *tmp; |
| |
| GST_DEBUG_OBJECT (demux, |
| "Recheck streams and sync to at least: %" GST_TIME_FORMAT, |
| GST_TIME_ARGS (time)); |
| |
| if (G_UNLIKELY (demux->program == NULL)) |
| return; |
| |
| /* Go over each stream and update it to at least 'time' time. |
| * For each stream, the pad stores the buffer counter the last time |
| * a gap check occurred (gap_ref_buffers) and a gap_ref_pts timestamp |
| * that is either the PTS from the stream or the PCR the pad was updated |
| * to. |
| * |
| * We can check nb_out_buffers to see if any buffers were pushed since then. |
| * This means we can detect buffers passing without PTSes fine and still generate |
| * gaps. |
| * |
| * If there haven't been any buffers pushed on this stream since the last |
| * gap check, push a gap event updating to the indicated input PCR time |
| * and update the pad's tracking. |
| * |
| * If there have been buffers pushed, update the reference buffer count |
| * and but don't push a gap event |
| */ |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *ps = (TSDemuxStream *) tmp->data; |
| GST_DEBUG_OBJECT (ps->pad, |
| "0x%04x, PTS:%" GST_TIME_FORMAT " REFPTS:%" GST_TIME_FORMAT " Gap:%" |
| GST_TIME_FORMAT " nb_buffers: %d (ref:%d)", |
| ((MpegTSBaseStream *) ps)->pid, GST_TIME_ARGS (ps->pts), |
| GST_TIME_ARGS (ps->gap_ref_pts), |
| GST_TIME_ARGS (ps->pts - ps->gap_ref_pts), ps->nb_out_buffers, |
| ps->gap_ref_buffers); |
| if (ps->pad == NULL) |
| continue; |
| |
| if (ps->nb_out_buffers == ps->gap_ref_buffers && ps->gap_ref_pts != ps->pts) { |
| /* Do initial setup of pad if needed - segment etc */ |
| GST_DEBUG_OBJECT (ps->pad, |
| "Stream needs update. Pushing GAP event to TS %" GST_TIME_FORMAT, |
| GST_TIME_ARGS (time)); |
| if (G_UNLIKELY (ps->need_newsegment)) |
| calculate_and_push_newsegment (demux, ps, NULL); |
| |
| /* Now send gap event */ |
| gst_pad_push_event (ps->pad, gst_event_new_gap (time, 0)); |
| } |
| |
| /* Update GAP tracking vars so we don't re-check this stream for a while */ |
| ps->gap_ref_pts = time; |
| if (ps->pts != GST_CLOCK_TIME_NONE && ps->pts > time) |
| ps->gap_ref_pts = ps->pts; |
| ps->gap_ref_buffers = ps->nb_out_buffers; |
| } |
| } |
| |
| static GstBufferList * |
| parse_opus_access_unit (TSDemuxStream * stream) |
| { |
| GstByteReader reader; |
| GstBufferList *buffer_list = NULL; |
| |
| buffer_list = gst_buffer_list_new (); |
| gst_byte_reader_init (&reader, stream->data, stream->current_size); |
| |
| do { |
| GstBuffer *buffer; |
| guint16 id; |
| guint au_size = 0; |
| guint8 b; |
| gboolean start_trim_flag, end_trim_flag, control_extension_flag; |
| guint16 start_trim = 0, end_trim = 0; |
| guint8 *packet_data; |
| guint packet_size; |
| |
| if (!gst_byte_reader_get_uint16_be (&reader, &id)) |
| goto error; |
| |
| /* No control header */ |
| if ((id >> 5) != 0x3ff) |
| goto error; |
| |
| do { |
| if (!gst_byte_reader_get_uint8 (&reader, &b)) |
| goto error; |
| au_size += b; |
| } while (b == 0xff); |
| |
| start_trim_flag = (id >> 4) & 0x1; |
| end_trim_flag = (id >> 3) & 0x1; |
| control_extension_flag = (id >> 2) & 0x1; |
| |
| if (start_trim_flag) { |
| if (!gst_byte_reader_get_uint16_be (&reader, &start_trim)) |
| goto error; |
| } |
| |
| if (end_trim_flag) { |
| if (!gst_byte_reader_get_uint16_be (&reader, &end_trim)) |
| goto error; |
| } |
| |
| if (control_extension_flag) { |
| if (!gst_byte_reader_get_uint8 (&reader, &b)) |
| goto error; |
| |
| if (!gst_byte_reader_skip (&reader, b)) |
| goto error; |
| } |
| |
| packet_size = au_size; |
| |
| /* FIXME: this should be |
| * packet_size = au_size - gst_byte_reader_get_pos (&reader); |
| * but ffmpeg and the only available sample stream from obe.tv |
| * are not including the control header size in au_size |
| */ |
| if (gst_byte_reader_get_remaining (&reader) < packet_size) |
| goto error; |
| if (!gst_byte_reader_dup_data (&reader, packet_size, &packet_data)) |
| goto error; |
| |
| buffer = gst_buffer_new_wrapped (packet_data, packet_size); |
| |
| if (start_trim != 0 || end_trim != 0) { |
| gst_buffer_add_audio_clipping_meta (buffer, GST_FORMAT_DEFAULT, |
| start_trim, end_trim); |
| } |
| |
| gst_buffer_list_add (buffer_list, buffer); |
| } while (gst_byte_reader_get_remaining (&reader) > 0); |
| |
| g_free (stream->data); |
| stream->data = NULL; |
| stream->current_size = 0; |
| |
| return buffer_list; |
| |
| error: |
| { |
| GST_ERROR ("Failed to parse Opus access unit"); |
| g_free (stream->data); |
| stream->data = NULL; |
| stream->current_size = 0; |
| if (buffer_list) |
| gst_buffer_list_unref (buffer_list); |
| return NULL; |
| } |
| } |
| |
| /* interlaced mode is disabled at the moment */ |
| /*#define TSDEMUX_JP2K_SUPPORT_INTERLACE */ |
| static GstBuffer * |
| parse_jp2k_access_unit (TSDemuxStream * stream) |
| { |
| GstByteReader reader; |
| /* header tag */ |
| guint32 header_tag; |
| /* Framerate box */ |
| guint16 den G_GNUC_UNUSED; |
| guint16 num G_GNUC_UNUSED; |
| /* Maximum bitrate box */ |
| guint32 MaxBr G_GNUC_UNUSED; |
| guint32 AUF[2] = { 0, 0 }; |
| #ifdef TSDEMUX_JP2K_SUPPORT_INTERLACE |
| /* Field Coding Box */ |
| guint8 Fic G_GNUC_UNUSED = 1; |
| guint8 Fio G_GNUC_UNUSED = 0; |
| /* header size equals 38 for non-interlaced, and 48 for interlaced */ |
| guint header_size = stream->jp2kInfos.interlace ? 48 : 38; |
| #else |
| /* header size equals 38 for non-interlaced, and 48 for interlaced */ |
| guint header_size = 38; |
| #endif |
| /* Time Code box */ |
| guint32 HHMMSSFF G_GNUC_UNUSED; |
| /* Broadcast color box */ |
| guint8 CollC G_GNUC_UNUSED; |
| guint8 b G_GNUC_UNUSED; |
| |
| guint data_location; |
| GstBuffer *retbuf = NULL; |
| |
| if (stream->current_size < header_size) { |
| GST_ERROR_OBJECT (stream->pad, "Not enough data for header"); |
| goto error; |
| } |
| |
| gst_byte_reader_init (&reader, stream->data, stream->current_size); |
| |
| /* Check for the location of the jp2k magic */ |
| data_location = |
| gst_byte_reader_masked_scan_uint32 (&reader, 0xffffffff, 0xff4fff51, 0, |
| stream->current_size); |
| GST_DEBUG_OBJECT (stream->pad, "data location %d", data_location); |
| if (data_location == -1) { |
| GST_ERROR_OBJECT (stream->pad, "Stream does not contain jp2k magic header"); |
| goto error; |
| } |
| |
| /* Elementary stream header box 'elsm' == 0x656c736d */ |
| header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| if (header_tag != 0x656c736d) { |
| GST_ERROR_OBJECT (stream->pad, "Expected ELSM box but found box %x instead", |
| header_tag); |
| goto error; |
| } |
| /* Frame rate box 'frat' == 0x66726174 */ |
| header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| if (header_tag != 0x66726174) { |
| GST_ERROR_OBJECT (stream->pad, |
| "Expected frame rate box, but found box %x instead", header_tag); |
| goto error; |
| |
| } |
| den = gst_byte_reader_get_uint16_be_unchecked (&reader); |
| num = gst_byte_reader_get_uint16_be_unchecked (&reader); |
| /* Maximum bit rate box 'brat' == 0x62726174 */ |
| header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| if (header_tag != 0x62726174) { |
| GST_ERROR_OBJECT (stream->pad, "Expected brat box but read box %x instead", |
| header_tag); |
| goto error; |
| |
| } |
| MaxBr = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| AUF[0] = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| if (stream->jp2kInfos.interlace) { |
| #ifdef TSDEMUX_JP2K_SUPPORT_INTERLACE |
| AUF[1] = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| /* Field Coding Box 'fiel' == 0x6669656c */ |
| header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| if (header_tag != 0x6669656c) { |
| GST_ERROR_OBJECT (stream->pad, |
| "Expected Field Coding box but found box %x instead", header_tag); |
| goto error; |
| } |
| Fic = gst_byte_reader_get_uint8_unchecked (&reader); |
| Fio = gst_byte_reader_get_uint8_unchecked (&reader); |
| #else |
| GST_ERROR_OBJECT (stream->pad, "interlaced mode not supported"); |
| goto error; |
| #endif |
| } |
| |
| /* Time Code Box 'tcod' == 0x74636f64 */ |
| /* Some progressive streams might have a AUF[1] of value 0 present */ |
| header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| if (header_tag == 0 && !stream->jp2kInfos.interlace) { |
| AUF[1] = header_tag; |
| header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| /* Bump up header size and recheck */ |
| header_size += 4; |
| if (stream->current_size < header_size) { |
| GST_ERROR_OBJECT (stream->pad, "Not enough data for header"); |
| goto error; |
| } |
| } |
| if (header_tag != 0x74636f64) { |
| GST_ERROR_OBJECT (stream->pad, |
| "Expected Time code box but found %d box instead", header_tag); |
| goto error; |
| } |
| HHMMSSFF = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| /* Broadcast Color Box 'bcol' == 0x6263686c */ |
| header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); |
| if (header_tag != 0x62636f6c) { |
| GST_ERROR_OBJECT (stream->pad, |
| "Expected Broadcast color box but found %x box instead", header_tag); |
| goto error; |
| } |
| CollC = gst_byte_reader_get_uint8_unchecked (&reader); |
| b = gst_byte_reader_get_uint8_unchecked (&reader); |
| |
| /* Check if we have enough data to create a valid buffer */ |
| if ((stream->current_size - data_location) < (AUF[0] + AUF[1])) { |
| GST_ERROR ("Required size (%d) greater than remaining size in buffer (%d)", |
| AUF[0] + AUF[1], (stream->current_size - data_location)); |
| goto error; |
| } |
| |
| retbuf = gst_buffer_new_wrapped_full (0, stream->data, stream->current_size, |
| data_location, stream->current_size - data_location, |
| stream->data, g_free); |
| stream->data = NULL; |
| stream->current_size = 0; |
| return retbuf; |
| |
| error: |
| GST_ERROR ("Failed to parse JP2K access unit"); |
| g_free (stream->data); |
| stream->data = NULL; |
| stream->current_size = 0; |
| return NULL; |
| } |
| |
| static GstFlowReturn |
| gst_ts_demux_push_pending_data (GstTSDemux * demux, TSDemuxStream * stream, |
| MpegTSBaseProgram * target_program) |
| { |
| GstFlowReturn res = GST_FLOW_OK; |
| MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; |
| GstBuffer *buffer = NULL; |
| GstBufferList *buffer_list = NULL; |
| |
| |
| GST_DEBUG_OBJECT (stream->pad, |
| "stream:%p, pid:0x%04x stream_type:%d state:%d", stream, bs->pid, |
| bs->stream_type, stream->state); |
| |
| if (G_UNLIKELY (stream->data == NULL)) { |
| GST_LOG ("stream->data == NULL"); |
| goto beach; |
| } |
| |
| if (G_UNLIKELY (stream->state == PENDING_PACKET_EMPTY)) { |
| GST_LOG ("EMPTY: returning"); |
| goto beach; |
| } |
| |
| if (G_UNLIKELY (stream->state != PENDING_PACKET_BUFFER)) { |
| GST_LOG ("state:%d, returning", stream->state); |
| goto beach; |
| } |
| |
| if (G_UNLIKELY (demux->program == NULL)) { |
| GST_LOG_OBJECT (demux, "No program"); |
| g_free (stream->data); |
| goto beach; |
| } |
| |
| if (stream->needs_keyframe) { |
| MpegTSBase *base = (MpegTSBase *) demux; |
| |
| if ((gst_ts_demux_adjust_seek_offset_for_keyframe (stream, stream->data, |
| stream->current_size)) || demux->last_seek_offset == 0) { |
| GST_DEBUG_OBJECT (stream->pad, |
| "Got Keyframe, ready to go at %" GST_TIME_FORMAT, |
| GST_TIME_ARGS (stream->pts)); |
| |
| if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_PRIVATE_PES_PACKETS && |
| bs->registration_id == DRF_ID_OPUS) { |
| buffer_list = parse_opus_access_unit (stream); |
| if (!buffer_list) { |
| res = GST_FLOW_ERROR; |
| goto beach; |
| } |
| |
| if (gst_buffer_list_length (buffer_list) == 1) { |
| buffer = gst_buffer_ref (gst_buffer_list_get (buffer_list, 0)); |
| gst_buffer_list_unref (buffer_list); |
| buffer_list = NULL; |
| } |
| } else if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K) { |
| buffer = parse_jp2k_access_unit (stream); |
| if (!buffer) { |
| res = GST_FLOW_ERROR; |
| goto beach; |
| } |
| } else { |
| buffer = gst_buffer_new_wrapped (stream->data, stream->current_size); |
| } |
| |
| stream->seeked_pts = stream->pts; |
| stream->seeked_dts = stream->dts; |
| stream->needs_keyframe = FALSE; |
| } else { |
| base->seek_offset = demux->last_seek_offset - 200 * base->packetsize; |
| if (demux->last_seek_offset < 200 * base->packetsize) |
| base->seek_offset = 0; |
| demux->last_seek_offset = base->seek_offset; |
| mpegts_packetizer_flush (base->packetizer, FALSE); |
| base->mode = BASE_MODE_SEEKING; |
| |
| stream->continuity_counter = CONTINUITY_UNSET; |
| res = GST_FLOW_REWINDING; |
| g_free (stream->data); |
| goto beach; |
| } |
| } else { |
| if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_PRIVATE_PES_PACKETS && |
| bs->registration_id == DRF_ID_OPUS) { |
| buffer_list = parse_opus_access_unit (stream); |
| if (!buffer_list) { |
| res = GST_FLOW_ERROR; |
| goto beach; |
| } |
| |
| if (gst_buffer_list_length (buffer_list) == 1) { |
| buffer = gst_buffer_ref (gst_buffer_list_get (buffer_list, 0)); |
| gst_buffer_list_unref (buffer_list); |
| buffer_list = NULL; |
| } |
| } else if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K) { |
| buffer = parse_jp2k_access_unit (stream); |
| if (!buffer) { |
| res = GST_FLOW_ERROR; |
| goto beach; |
| } |
| } else { |
| buffer = gst_buffer_new_wrapped (stream->data, stream->current_size); |
| } |
| |
| if (G_UNLIKELY (stream->pending_ts && !check_pending_buffers (demux))) { |
| if (buffer) { |
| PendingBuffer *pend; |
| pend = g_slice_new0 (PendingBuffer); |
| pend->buffer = buffer; |
| pend->pts = stream->raw_pts; |
| pend->dts = stream->raw_dts; |
| stream->pending = g_list_append (stream->pending, pend); |
| } else { |
| guint i, n; |
| |
| n = gst_buffer_list_length (buffer_list); |
| for (i = 0; i < n; i++) { |
| PendingBuffer *pend; |
| pend = g_slice_new0 (PendingBuffer); |
| pend->buffer = gst_buffer_ref (gst_buffer_list_get (buffer_list, i)); |
| pend->pts = i == 0 ? stream->raw_pts : -1; |
| pend->dts = i == 0 ? stream->raw_dts : -1; |
| stream->pending = g_list_append (stream->pending, pend); |
| } |
| gst_buffer_list_unref (buffer_list); |
| } |
| GST_DEBUG ("Not enough information to push buffers yet, storing buffer"); |
| goto beach; |
| } |
| } |
| |
| if (G_UNLIKELY (stream->need_newsegment)) |
| calculate_and_push_newsegment (demux, stream, target_program); |
| |
| /* FIXME : Push pending buffers if any */ |
| if (G_UNLIKELY (stream->pending)) { |
| GList *tmp; |
| for (tmp = stream->pending; tmp; tmp = tmp->next) { |
| PendingBuffer *pend = (PendingBuffer *) tmp->data; |
| |
| GST_DEBUG_OBJECT (stream->pad, |
| "Pushing pending buffer PTS:%" GST_TIME_FORMAT " DTS:%" |
| GST_TIME_FORMAT, GST_TIME_ARGS (GST_BUFFER_PTS (pend->buffer)), |
| GST_TIME_ARGS (GST_BUFFER_DTS (pend->buffer))); |
| |
| if (stream->discont) |
| GST_BUFFER_FLAG_SET (pend->buffer, GST_BUFFER_FLAG_DISCONT); |
| stream->discont = FALSE; |
| |
| res = gst_pad_push (stream->pad, pend->buffer); |
| stream->nb_out_buffers += 1; |
| g_slice_free (PendingBuffer, pend); |
| } |
| g_list_free (stream->pending); |
| stream->pending = NULL; |
| } |
| |
| if ((GST_CLOCK_TIME_IS_VALID (stream->seeked_pts) |
| && stream->pts < stream->seeked_pts) || |
| (GST_CLOCK_TIME_IS_VALID (stream->seeked_dts) && |
| stream->pts < stream->seeked_dts)) { |
| GST_INFO_OBJECT (stream->pad, |
| "Droping with PTS: %" GST_TIME_FORMAT " DTS: %" GST_TIME_FORMAT |
| " after seeking as other stream needed to be seeked further" |
| "(seeked PTS: %" GST_TIME_FORMAT " DTS: %" GST_TIME_FORMAT ")", |
| GST_TIME_ARGS (stream->pts), GST_TIME_ARGS (stream->dts), |
| GST_TIME_ARGS (stream->seeked_pts), GST_TIME_ARGS (stream->seeked_dts)); |
| if (buffer) |
| gst_buffer_unref (buffer); |
| if (buffer_list) |
| gst_buffer_list_unref (buffer_list); |
| goto beach; |
| } |
| |
| GST_DEBUG_OBJECT (stream->pad, "stream->pts %" GST_TIME_FORMAT, |
| GST_TIME_ARGS (stream->pts)); |
| |
| /* Decorate buffer or first buffer of the buffer list */ |
| if (buffer_list) |
| buffer = gst_buffer_list_get (buffer_list, 0); |
| |
| if (GST_CLOCK_TIME_IS_VALID (stream->pts)) |
| GST_BUFFER_PTS (buffer) = stream->pts; |
| if (GST_CLOCK_TIME_IS_VALID (stream->dts)) |
| GST_BUFFER_DTS (buffer) = stream->dts; |
| |
| if (stream->discont) |
| GST_BUFFER_FLAG_SET (buffer, GST_BUFFER_FLAG_DISCONT); |
| stream->discont = FALSE; |
| |
| if (buffer_list) |
| buffer = NULL; |
| |
| GST_DEBUG_OBJECT (stream->pad, |
| "Pushing buffer%s with PTS: %" GST_TIME_FORMAT " , DTS: %" |
| GST_TIME_FORMAT, (buffer_list ? "list" : ""), GST_TIME_ARGS (stream->pts), |
| GST_TIME_ARGS (stream->dts)); |
| |
| if (GST_CLOCK_TIME_IS_VALID (stream->dts)) |
| demux->segment.position = stream->dts; |
| else if (GST_CLOCK_TIME_IS_VALID (stream->pts)) |
| demux->segment.position = stream->pts; |
| |
| if (buffer) { |
| res = gst_pad_push (stream->pad, buffer); |
| /* Record that a buffer was pushed */ |
| stream->nb_out_buffers += 1; |
| } else { |
| guint n = gst_buffer_list_length (buffer_list); |
| res = gst_pad_push_list (stream->pad, buffer_list); |
| /* Record that a buffer was pushed */ |
| stream->nb_out_buffers += n; |
| } |
| GST_DEBUG_OBJECT (stream->pad, "Returned %s", gst_flow_get_name (res)); |
| res = gst_flow_combiner_update_flow (demux->flowcombiner, res); |
| GST_DEBUG_OBJECT (stream->pad, "combined %s", gst_flow_get_name (res)); |
| |
| /* GAP / sparse stream tracking */ |
| if (G_UNLIKELY (stream->gap_ref_pts == GST_CLOCK_TIME_NONE)) |
| stream->gap_ref_pts = stream->pts; |
| else { |
| /* Look if the stream PTS has advanced 2 seconds since the last |
| * gap check, and sync streams if it has. The first stream to |
| * hit this will trigger a gap check */ |
| if (G_UNLIKELY (stream->pts != GST_CLOCK_TIME_NONE && |
| stream->pts > stream->gap_ref_pts + 2 * GST_SECOND)) { |
| if (demux->program->pcr_pid != 0x1fff) { |
| GstClockTime curpcr = |
| mpegts_packetizer_get_current_time (MPEG_TS_BASE_PACKETIZER (demux), |
| demux->program->pcr_pid); |
| if (curpcr == GST_CLOCK_TIME_NONE || curpcr < 800 * GST_MSECOND) |
| goto beach; |
| curpcr -= 800 * GST_MSECOND; |
| /* Use the current PCR (with a safety margin) to sync against */ |
| gst_ts_demux_check_and_sync_streams (demux, curpcr); |
| } else { |
| /* If we don't have a PCR track, just use the current stream PTS */ |
| gst_ts_demux_check_and_sync_streams (demux, stream->pts); |
| } |
| } |
| } |
| |
| beach: |
| /* Reset everything */ |
| GST_LOG ("Resetting to EMPTY, returning %s", gst_flow_get_name (res)); |
| stream->state = PENDING_PACKET_EMPTY; |
| stream->data = NULL; |
| stream->expected_size = 0; |
| stream->current_size = 0; |
| |
| return res; |
| } |
| |
| static GstFlowReturn |
| gst_ts_demux_handle_packet (GstTSDemux * demux, TSDemuxStream * stream, |
| MpegTSPacketizerPacket * packet, GstMpegtsSection * section) |
| { |
| GstFlowReturn res = GST_FLOW_OK; |
| |
| GST_LOG ("pid 0x%04x pusi:%d, afc:%d, cont:%d, payload:%p", packet->pid, |
| packet->payload_unit_start_indicator, packet->scram_afc_cc & 0x30, |
| FLAGS_CONTINUITY_COUNTER (packet->scram_afc_cc), packet->payload); |
| |
| if (G_UNLIKELY (packet->payload_unit_start_indicator) && |
| FLAGS_HAS_PAYLOAD (packet->scram_afc_cc)) |
| /* Flush previous data */ |
| res = gst_ts_demux_push_pending_data (demux, stream, NULL); |
| |
| if (packet->payload && (res == GST_FLOW_OK || res == GST_FLOW_NOT_LINKED) |
| && stream->pad) { |
| gst_ts_demux_queue_data (demux, stream, packet); |
| GST_LOG ("current_size:%d, expected_size:%d", |
| stream->current_size, stream->expected_size); |
| /* Finally check if the data we queued completes a packet */ |
| if (stream->expected_size && stream->current_size == stream->expected_size) { |
| GST_LOG ("pushing complete packet"); |
| res = gst_ts_demux_push_pending_data (demux, stream, NULL); |
| } |
| } |
| |
| /* We are rewinding to find a keyframe, |
| * and didn't want the data to be queued |
| */ |
| if (res == GST_FLOW_REWINDING) |
| res = GST_FLOW_OK; |
| |
| return res; |
| } |
| |
| static void |
| gst_ts_demux_flush (MpegTSBase * base, gboolean hard) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX_CAST (base); |
| |
| gst_ts_demux_flush_streams (demux, hard); |
| |
| if (demux->segment_event) { |
| gst_event_unref (demux->segment_event); |
| demux->segment_event = NULL; |
| } |
| if (demux->global_tags) { |
| gst_tag_list_unref (demux->global_tags); |
| demux->global_tags = NULL; |
| } |
| if (hard) { |
| /* For pull mode seeks the current segment needs to be preserved */ |
| demux->rate = 1.0; |
| gst_segment_init (&demux->segment, GST_FORMAT_UNDEFINED); |
| } |
| } |
| |
| static GstFlowReturn |
| gst_ts_demux_drain (MpegTSBase * base) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX_CAST (base); |
| GList *tmp; |
| GstFlowReturn res = GST_FLOW_OK; |
| |
| if (!demux->program) |
| return res; |
| |
| for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { |
| TSDemuxStream *stream = (TSDemuxStream *) tmp->data; |
| if (stream->pad) { |
| res = gst_ts_demux_push_pending_data (demux, stream, NULL); |
| if (G_UNLIKELY (res != GST_FLOW_OK)) |
| break; |
| } |
| } |
| |
| return res; |
| } |
| |
| static GstFlowReturn |
| gst_ts_demux_push (MpegTSBase * base, MpegTSPacketizerPacket * packet, |
| GstMpegtsSection * section) |
| { |
| GstTSDemux *demux = GST_TS_DEMUX_CAST (base); |
| TSDemuxStream *stream = NULL; |
| GstFlowReturn res = GST_FLOW_OK; |
| |
| if (G_LIKELY (demux->program)) { |
| stream = (TSDemuxStream *) demux->program->streams[packet->pid]; |
| |
| if (stream) { |
| res = gst_ts_demux_handle_packet (demux, stream, packet, section); |
| } |
| } |
| return res; |
| } |
| |
| gboolean |
| gst_ts_demux_plugin_init (GstPlugin * plugin) |
| { |
| GST_DEBUG_CATEGORY_INIT (ts_demux_debug, "tsdemux", 0, |
| "MPEG transport stream demuxer"); |
| init_pes_parser (); |
| |
| return gst_element_register (plugin, "tsdemux", |
| GST_RANK_PRIMARY, GST_TYPE_TS_DEMUX); |
| } |