/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this file, * You can obtain one at http://mozilla.org/MPL/2.0/. */ #if !defined(GStreamerReader_h_) #define GStreamerReader_h_ #include #include #include #include #include "MediaDecoderReader.h" namespace mozilla { namespace dom { class TimeRanges; } class AbstractMediaDecoder; class GStreamerReader : public MediaDecoderReader { public: GStreamerReader(AbstractMediaDecoder* aDecoder); virtual ~GStreamerReader(); virtual nsresult Init(MediaDecoderReader* aCloneDonor); virtual nsresult ResetDecode(); virtual bool DecodeAudioData(); virtual bool DecodeVideoFrame(bool &aKeyframeSkip, int64_t aTimeThreshold); virtual nsresult ReadMetadata(VideoInfo* aInfo, MetadataTags** aTags); virtual nsresult Seek(int64_t aTime, int64_t aStartTime, int64_t aEndTime, int64_t aCurrentTime); virtual nsresult GetBuffered(dom::TimeRanges* aBuffered, int64_t aStartTime); virtual bool HasAudio() { return mInfo.mHasAudio; } virtual bool HasVideo() { return mInfo.mHasVideo; } private: void ReadAndPushData(guint aLength); bool WaitForDecodedData(int *counter); void NotifyBytesConsumed(); int64_t QueryDuration(); /* Gst callbacks */ /* Called on the source-setup signal emitted by playbin. Used to * configure appsrc . */ static void PlayBinSourceSetupCb(GstElement *aPlayBin, GParamSpec *pspec, gpointer aUserData); void PlayBinSourceSetup(GstAppSrc *aSource); /* Called from appsrc when we need to read more data from the resource */ static void NeedDataCb(GstAppSrc *aSrc, guint aLength, gpointer aUserData); void NeedData(GstAppSrc *aSrc, guint aLength); /* Called when appsrc has enough data and we can stop reading */ static void EnoughDataCb(GstAppSrc *aSrc, gpointer aUserData); void EnoughData(GstAppSrc *aSrc); /* Called when a seek is issued on the pipeline */ static gboolean SeekDataCb(GstAppSrc *aSrc, guint64 aOffset, gpointer aUserData); gboolean SeekData(GstAppSrc *aSrc, guint64 aOffset); /* Called when events reach the sinks. See inline comments */ static gboolean EventProbeCb(GstPad *aPad, GstEvent *aEvent, gpointer aUserData); gboolean EventProbe(GstPad *aPad, GstEvent *aEvent); /* Called when the pipeline is prerolled, that is when at start or after a * seek, the first audio and video buffers are queued in the sinks. */ static GstFlowReturn NewPrerollCb(GstAppSink *aSink, gpointer aUserData); void VideoPreroll(); void AudioPreroll(); /* Called when buffers reach the sinks */ static GstFlowReturn NewBufferCb(GstAppSink *aSink, gpointer aUserData); void NewVideoBuffer(); void NewAudioBuffer(); /* Called at end of stream, when decoding has finished */ static void EosCb(GstAppSink *aSink, gpointer aUserData); void Eos(GstAppSink *aSink); GstElement *mPlayBin; GstBus *mBus; GstAppSrc *mSource; /* video sink bin */ GstElement *mVideoSink; /* the actual video app sink */ GstAppSink *mVideoAppSink; /* audio sink bin */ GstElement *mAudioSink; /* the actual audio app sink */ GstAppSink *mAudioAppSink; GstVideoFormat mFormat; nsIntRect mPicture; int mVideoSinkBufferCount; int mAudioSinkBufferCount; GstAppSrcCallbacks mSrcCallbacks; GstAppSinkCallbacks mSinkCallbacks; /* monitor used to synchronize access to shared state between gstreamer * threads and other gecko threads */ ReentrantMonitor mGstThreadsMonitor; /* video and audio segments we use to convert absolute timestamps to [0, * stream_duration]. They're set when the pipeline is started or after a seek. * Concurrent access guarded with mGstThreadsMonitor. */ GstSegment mVideoSegment; GstSegment mAudioSegment; /* bool used to signal when gst has detected the end of stream and * DecodeAudioData and DecodeVideoFrame should not expect any more data */ bool mReachedEos; /* offset we've reached reading from the source */ gint64 mByteOffset; /* the last offset we reported with NotifyBytesConsumed */ gint64 mLastReportedByteOffset; int fpsNum; int fpsDen; }; } // namespace mozilla #endif