/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this file, * You can obtain one at http://mozilla.org/MPL/2.0/. */ #if !defined(nsGStreamerReader_h_) #define nsGStreamerReader_h_ #include #include #include #include #include "nsBuiltinDecoderReader.h" using namespace mozilla; class nsMediaDecoder; class nsTimeRanges; class nsGStreamerReader : public nsBuiltinDecoderReader { public: nsGStreamerReader(nsBuiltinDecoder* aDecoder); virtual ~nsGStreamerReader(); virtual nsresult Init(nsBuiltinDecoderReader* aCloneDonor); virtual nsresult ResetDecode(); virtual bool DecodeAudioData(); virtual bool DecodeVideoFrame(bool &aKeyframeSkip, PRInt64 aTimeThreshold); virtual nsresult ReadMetadata(nsVideoInfo* aInfo); virtual nsresult Seek(PRInt64 aTime, PRInt64 aStartTime, PRInt64 aEndTime, PRInt64 aCurrentTime); virtual nsresult GetBuffered(nsTimeRanges* aBuffered, PRInt64 aStartTime); virtual bool HasAudio() { return mInfo.mHasAudio; } virtual bool HasVideo() { return mInfo.mHasVideo; } private: void ReadAndPushData(guint aLength); bool WaitForDecodedData(int *counter); void NotifyBytesConsumed(); PRInt64 QueryDuration(); /* Gst callbacks */ /* Called on the source-setup signal emitted by playbin. Used to * configure appsrc . */ static void PlayBinSourceSetupCb(GstElement *aPlayBin, GstElement *aSource, gpointer aUserData); void PlayBinSourceSetup(GstAppSrc *aSource); /* Called from appsrc when we need to read more data from the resource */ static void NeedDataCb(GstAppSrc *aSrc, guint aLength, gpointer aUserData); void NeedData(GstAppSrc *aSrc, guint aLength); /* Called when appsrc has enough data and we can stop reading */ static void EnoughDataCb(GstAppSrc *aSrc, gpointer aUserData); void EnoughData(GstAppSrc *aSrc); /* Called when a seek is issued on the pipeline */ static gboolean SeekDataCb(GstAppSrc *aSrc, guint64 aOffset, gpointer aUserData); gboolean SeekData(GstAppSrc *aSrc, guint64 aOffset); /* Called when events reach the sinks. See inline comments */ static gboolean EventProbeCb(GstPad *aPad, GstEvent *aEvent, gpointer aUserData); gboolean EventProbe(GstPad *aPad, GstEvent *aEvent); /* Called when the pipeline is prerolled, that is when at start or after a * seek, the first audio and video buffers are queued in the sinks. */ static GstFlowReturn NewPrerollCb(GstAppSink *aSink, gpointer aUserData); void VideoPreroll(); void AudioPreroll(); /* Called when buffers reach the sinks */ static GstFlowReturn NewBufferCb(GstAppSink *aSink, gpointer aUserData); void NewVideoBuffer(); void NewAudioBuffer(); /* Called at end of stream, when decoding has finished */ static void EosCb(GstAppSink *aSink, gpointer aUserData); void Eos(GstAppSink *aSink); GstElement *mPlayBin; GstBus *mBus; GstAppSrc *mSource; /* video sink bin */ GstElement *mVideoSink; /* the actual video app sink */ GstAppSink *mVideoAppSink; /* audio sink bin */ GstElement *mAudioSink; /* the actual audio app sink */ GstAppSink *mAudioAppSink; GstVideoFormat mFormat; nsIntRect mPicture; int mVideoSinkBufferCount; int mAudioSinkBufferCount; GstAppSrcCallbacks mSrcCallbacks; GstAppSinkCallbacks mSinkCallbacks; /* monitor used to synchronize access to shared state between gstreamer * threads and other gecko threads */ mozilla::ReentrantMonitor mGstThreadsMonitor; /* video and audio segments we use to convert absolute timestamps to [0, * stream_duration]. They're set when the pipeline is started or after a seek. * Concurrent access guarded with mGstThreadsMonitor. */ GstSegment mVideoSegment; GstSegment mAudioSegment; /* bool used to signal when gst has detected the end of stream and * DecodeAudioData and DecodeVideoFrame should not expect any more data */ bool mReachedEos; /* offset we've reached reading from the source */ gint64 mByteOffset; /* the last offset we reported with NotifyBytesConsumed */ gint64 mLastReportedByteOffset; int fpsNum; int fpsDen; }; #endif