mirror of
https://gitlab.winehq.org/wine/wine-gecko.git
synced 2024-09-13 09:24:08 -07:00
152 lines
4.2 KiB
C++
152 lines
4.2 KiB
C++
#include <OMX.h>
|
|
#include <stagefright/MediaSource.h>
|
|
#include <stagefright/DataSource.h>
|
|
|
|
#include <utils/RefBase.h>
|
|
|
|
#include "GonkNativeWindow.h"
|
|
#include "GonkIOSurfaceImage.h"
|
|
#include "MPAPI.h"
|
|
#include "MediaResource.h"
|
|
#include "nsBuiltinDecoder.h"
|
|
|
|
namespace mozilla {
|
|
namespace layers {
|
|
|
|
class VideoGraphicBuffer : public GraphicBufferLocked {
|
|
// XXX change this to an actual smart pointer at some point
|
|
android::MediaBuffer *mMediaBuffer;
|
|
public:
|
|
VideoGraphicBuffer(android::MediaBuffer *aBuffer,
|
|
SurfaceDescriptor *aDescriptor);
|
|
~VideoGraphicBuffer();
|
|
void Unlock();
|
|
};
|
|
|
|
}
|
|
}
|
|
|
|
namespace android {
|
|
|
|
// MediaStreamSource is a DataSource that reads from a MPAPI media stream.
|
|
class MediaStreamSource : public DataSource {
|
|
typedef mozilla::MediaResource MediaResource;
|
|
|
|
MediaResource *mResource;
|
|
nsBuiltinDecoder *mDecoder;
|
|
public:
|
|
MediaStreamSource(MediaResource *aResource,
|
|
nsBuiltinDecoder *aDecoder);
|
|
|
|
virtual status_t initCheck() const;
|
|
virtual ssize_t readAt(off64_t offset, void *data, size_t size);
|
|
virtual ssize_t readAt(off_t offset, void *data, size_t size) {
|
|
return readAt(static_cast<off64_t>(offset), data, size);
|
|
}
|
|
virtual status_t getSize(off_t *size) {
|
|
off64_t size64;
|
|
status_t status = getSize(&size64);
|
|
*size = size64;
|
|
return status;
|
|
}
|
|
virtual status_t getSize(off64_t *size);
|
|
virtual uint32_t flags() {
|
|
return kWantsPrefetching;
|
|
}
|
|
|
|
virtual ~MediaStreamSource();
|
|
|
|
private:
|
|
MediaStreamSource(const MediaStreamSource &);
|
|
MediaStreamSource &operator=(const MediaStreamSource &);
|
|
};
|
|
|
|
class OmxDecoder {
|
|
typedef MPAPI::AudioFrame AudioFrame;
|
|
typedef MPAPI::VideoFrame VideoFrame;
|
|
typedef mozilla::MediaResource MediaResource;
|
|
|
|
enum {
|
|
kPreferSoftwareCodecs = 1
|
|
};
|
|
|
|
nsBuiltinDecoder *mDecoder;
|
|
MediaResource *mResource;
|
|
sp<GonkNativeWindow> mNativeWindow;
|
|
sp<MediaSource> mVideoTrack;
|
|
sp<MediaSource> mVideoSource;
|
|
sp<MediaSource> mAudioTrack;
|
|
sp<MediaSource> mAudioSource;
|
|
int32_t mVideoWidth;
|
|
int32_t mVideoHeight;
|
|
int32_t mVideoColorFormat;
|
|
int32_t mVideoStride;
|
|
int32_t mVideoSliceHeight;
|
|
int32_t mVideoRotation;
|
|
int32_t mAudioChannels;
|
|
int32_t mAudioSampleRate;
|
|
int64_t mDurationUs;
|
|
VideoFrame mVideoFrame;
|
|
AudioFrame mAudioFrame;
|
|
|
|
// Lifetime of these should be handled by OMXCodec, as long as we release
|
|
// them after use: see ReleaseVideoBuffer(), ReleaseAudioBuffer()
|
|
MediaBuffer *mVideoBuffer;
|
|
MediaBuffer *mAudioBuffer;
|
|
|
|
// 'true' if a read from the audio stream was done while reading the metadata
|
|
bool mAudioMetadataRead;
|
|
|
|
void ReleaseVideoBuffer();
|
|
void ReleaseAudioBuffer();
|
|
|
|
void PlanarYUV420Frame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame);
|
|
void CbYCrYFrame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame);
|
|
void SemiPlanarYUV420Frame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame);
|
|
void SemiPlanarYVU420Frame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame);
|
|
bool ToVideoFrame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame);
|
|
bool ToAudioFrame(AudioFrame *aFrame, int64_t aTimeUs, void *aData, size_t aDataOffset, size_t aSize,
|
|
int32_t aAudioChannels, int32_t aAudioSampleRate);
|
|
public:
|
|
OmxDecoder(MediaResource *aResource, nsBuiltinDecoder *aDecoder);
|
|
~OmxDecoder();
|
|
|
|
bool Init();
|
|
bool SetVideoFormat();
|
|
bool SetAudioFormat();
|
|
|
|
void GetDuration(int64_t *durationUs) {
|
|
*durationUs = mDurationUs;
|
|
}
|
|
|
|
void GetVideoParameters(int32_t *width, int32_t *height) {
|
|
*width = mVideoWidth;
|
|
*height = mVideoHeight;
|
|
}
|
|
|
|
void GetAudioParameters(int32_t *numChannels, int32_t *sampleRate) {
|
|
*numChannels = mAudioChannels;
|
|
*sampleRate = mAudioSampleRate;
|
|
}
|
|
|
|
bool HasVideo() {
|
|
return mVideoSource != nullptr;
|
|
}
|
|
|
|
bool HasAudio() {
|
|
return mAudioSource != nullptr;
|
|
}
|
|
|
|
bool ReadVideo(VideoFrame *aFrame, int64_t aSeekTimeUs,
|
|
bool aKeyframeSkip = false,
|
|
bool aDoSeek = false);
|
|
bool ReadAudio(AudioFrame *aFrame, int64_t aSeekTimeUs);
|
|
|
|
MediaResource *GetResource() {
|
|
return mResource;
|
|
}
|
|
};
|
|
|
|
}
|
|
|