mirror of
https://gitlab.winehq.org/wine/wine-gecko.git
synced 2024-09-13 09:24:08 -07:00
162 lines
5.5 KiB
C++
162 lines
5.5 KiB
C++
/* This Source Code Form is subject to the terms of the Mozilla Public
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this file,
|
|
* You can obtain one at http://mozilla.org/MPL/2.0/. */
|
|
|
|
#if !defined(GStreamerReader_h_)
|
|
#define GStreamerReader_h_
|
|
|
|
#include <gst/gst.h>
|
|
#include <gst/app/gstappsrc.h>
|
|
#include <gst/app/gstappsink.h>
|
|
#pragma GCC diagnostic push
|
|
#pragma GCC diagnostic ignored "-Wpragmas"
|
|
#pragma GCC diagnostic ignored "-Wreserved-user-defined-literal"
|
|
#include <gst/video/video.h>
|
|
#pragma GCC diagnostic pop
|
|
#include <map>
|
|
#include "MediaDecoderReader.h"
|
|
|
|
namespace mozilla {
|
|
|
|
namespace dom {
|
|
class TimeRanges;
|
|
}
|
|
|
|
class AbstractMediaDecoder;
|
|
|
|
class GStreamerReader : public MediaDecoderReader
|
|
{
|
|
public:
|
|
GStreamerReader(AbstractMediaDecoder* aDecoder);
|
|
virtual ~GStreamerReader();
|
|
|
|
virtual nsresult Init(MediaDecoderReader* aCloneDonor);
|
|
virtual nsresult ResetDecode();
|
|
virtual bool DecodeAudioData();
|
|
virtual bool DecodeVideoFrame(bool &aKeyframeSkip,
|
|
int64_t aTimeThreshold);
|
|
virtual nsresult ReadMetadata(VideoInfo* aInfo,
|
|
MetadataTags** aTags);
|
|
virtual nsresult Seek(int64_t aTime,
|
|
int64_t aStartTime,
|
|
int64_t aEndTime,
|
|
int64_t aCurrentTime);
|
|
virtual nsresult GetBuffered(dom::TimeRanges* aBuffered, int64_t aStartTime);
|
|
|
|
virtual bool HasAudio() {
|
|
return mInfo.mHasAudio;
|
|
}
|
|
|
|
virtual bool HasVideo() {
|
|
return mInfo.mHasVideo;
|
|
}
|
|
|
|
private:
|
|
|
|
void ReadAndPushData(guint aLength);
|
|
void NotifyBytesConsumed();
|
|
int64_t QueryDuration();
|
|
|
|
/* Called once the pipeline is setup to check that the stream only contains
|
|
* supported formats
|
|
*/
|
|
nsresult CheckSupportedFormats();
|
|
|
|
/* Gst callbacks */
|
|
|
|
/* Called on the source-setup signal emitted by playbin. Used to
|
|
* configure appsrc .
|
|
*/
|
|
static void PlayBinSourceSetupCb(GstElement* aPlayBin,
|
|
GParamSpec* pspec,
|
|
gpointer aUserData);
|
|
void PlayBinSourceSetup(GstAppSrc* aSource);
|
|
|
|
/* Called from appsrc when we need to read more data from the resource */
|
|
static void NeedDataCb(GstAppSrc* aSrc, guint aLength, gpointer aUserData);
|
|
void NeedData(GstAppSrc* aSrc, guint aLength);
|
|
|
|
/* Called when appsrc has enough data and we can stop reading */
|
|
static void EnoughDataCb(GstAppSrc* aSrc, gpointer aUserData);
|
|
void EnoughData(GstAppSrc* aSrc);
|
|
|
|
/* Called when a seek is issued on the pipeline */
|
|
static gboolean SeekDataCb(GstAppSrc* aSrc,
|
|
guint64 aOffset,
|
|
gpointer aUserData);
|
|
gboolean SeekData(GstAppSrc* aSrc, guint64 aOffset);
|
|
|
|
/* Called when events reach the sinks. See inline comments */
|
|
static gboolean EventProbeCb(GstPad* aPad, GstEvent* aEvent, gpointer aUserData);
|
|
gboolean EventProbe(GstPad* aPad, GstEvent* aEvent);
|
|
|
|
/* Called when elements in the video branch of the pipeline call
|
|
* gst_pad_alloc_buffer(). Used to provide PlanarYCbCrImage backed GstBuffers
|
|
* to the pipeline so that a memory copy can be avoided when handling YUV
|
|
* buffers from the pipeline to the gfx side.
|
|
*/
|
|
static GstFlowReturn AllocateVideoBufferCb(GstPad* aPad, guint64 aOffset, guint aSize,
|
|
GstCaps* aCaps, GstBuffer** aBuf);
|
|
GstFlowReturn AllocateVideoBufferFull(GstPad* aPad, guint64 aOffset, guint aSize,
|
|
GstCaps* aCaps, GstBuffer** aBuf, nsRefPtr<layers::PlanarYCbCrImage>& aImage);
|
|
GstFlowReturn AllocateVideoBuffer(GstPad* aPad, guint64 aOffset, guint aSize,
|
|
GstCaps* aCaps, GstBuffer** aBuf);
|
|
|
|
/* Called when the pipeline is prerolled, that is when at start or after a
|
|
* seek, the first audio and video buffers are queued in the sinks.
|
|
*/
|
|
static GstFlowReturn NewPrerollCb(GstAppSink* aSink, gpointer aUserData);
|
|
void VideoPreroll();
|
|
void AudioPreroll();
|
|
|
|
/* Called when buffers reach the sinks */
|
|
static GstFlowReturn NewBufferCb(GstAppSink* aSink, gpointer aUserData);
|
|
void NewVideoBuffer();
|
|
void NewAudioBuffer();
|
|
|
|
/* Called at end of stream, when decoding has finished */
|
|
static void EosCb(GstAppSink* aSink, gpointer aUserData);
|
|
void Eos(GstAppSink* aSink);
|
|
|
|
GstElement* mPlayBin;
|
|
GstBus* mBus;
|
|
GstAppSrc* mSource;
|
|
/* video sink bin */
|
|
GstElement* mVideoSink;
|
|
/* the actual video app sink */
|
|
GstAppSink* mVideoAppSink;
|
|
/* audio sink bin */
|
|
GstElement* mAudioSink;
|
|
/* the actual audio app sink */
|
|
GstAppSink* mAudioAppSink;
|
|
GstVideoFormat mFormat;
|
|
nsIntRect mPicture;
|
|
int mVideoSinkBufferCount;
|
|
int mAudioSinkBufferCount;
|
|
GstAppSrcCallbacks mSrcCallbacks;
|
|
GstAppSinkCallbacks mSinkCallbacks;
|
|
/* monitor used to synchronize access to shared state between gstreamer
|
|
* threads and other gecko threads */
|
|
ReentrantMonitor mGstThreadsMonitor;
|
|
/* video and audio segments we use to convert absolute timestamps to [0,
|
|
* stream_duration]. They're set when the pipeline is started or after a seek.
|
|
* Concurrent access guarded with mGstThreadsMonitor.
|
|
*/
|
|
GstSegment mVideoSegment;
|
|
GstSegment mAudioSegment;
|
|
/* bool used to signal when gst has detected the end of stream and
|
|
* DecodeAudioData and DecodeVideoFrame should not expect any more data
|
|
*/
|
|
bool mReachedEos;
|
|
/* offset we've reached reading from the source */
|
|
gint64 mByteOffset;
|
|
/* the last offset we reported with NotifyBytesConsumed */
|
|
gint64 mLastReportedByteOffset;
|
|
int fpsNum;
|
|
int fpsDen;
|
|
};
|
|
|
|
} // namespace mozilla
|
|
|
|
#endif
|