2012-09-26 21:33:43 -07:00
|
|
|
/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
|
|
|
|
/* vim:set ts=2 sw=2 sts=2 et cindent: */
|
|
|
|
/* This Source Code Form is subject to the terms of the Mozilla Public
|
|
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this file,
|
|
|
|
* You can obtain one at http://mozilla.org/MPL/2.0/. */
|
|
|
|
|
|
|
|
#include "nsMediaOmxReader.h"
|
|
|
|
|
|
|
|
#include "mozilla/TimeStamp.h"
|
|
|
|
#include "nsTimeRanges.h"
|
|
|
|
#include "MediaResource.h"
|
|
|
|
#include "VideoUtils.h"
|
|
|
|
#include "nsMediaOmxDecoder.h"
|
|
|
|
|
|
|
|
using namespace android;
|
|
|
|
using namespace mozilla;
|
|
|
|
|
|
|
|
nsMediaOmxReader::nsMediaOmxReader(nsBuiltinDecoder *aDecoder) :
|
|
|
|
nsBuiltinDecoderReader(aDecoder),
|
|
|
|
mOmxDecoder(nullptr),
|
|
|
|
mHasVideo(false),
|
|
|
|
mHasAudio(false),
|
|
|
|
mVideoSeekTimeUs(-1),
|
|
|
|
mAudioSeekTimeUs(-1),
|
|
|
|
mLastVideoFrame(nullptr)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
nsMediaOmxReader::~nsMediaOmxReader()
|
|
|
|
{
|
|
|
|
ResetDecode();
|
|
|
|
}
|
|
|
|
|
|
|
|
nsresult nsMediaOmxReader::Init(nsBuiltinDecoderReader* aCloneDonor)
|
|
|
|
{
|
|
|
|
return NS_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
nsresult nsMediaOmxReader::ReadMetadata(nsVideoInfo* aInfo,
|
|
|
|
nsHTMLMediaElement::MetadataTags** aTags)
|
|
|
|
{
|
|
|
|
NS_ASSERTION(mDecoder->OnDecodeThread(), "Should be on decode thread.");
|
|
|
|
|
|
|
|
*aTags = nullptr;
|
|
|
|
|
|
|
|
if (!mOmxDecoder) {
|
|
|
|
mOmxDecoder = new OmxDecoder(mDecoder->GetResource(), mDecoder);
|
|
|
|
mOmxDecoder->Init();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set the total duration (the max of the audio and video track).
|
|
|
|
int64_t durationUs;
|
|
|
|
mOmxDecoder->GetDuration(&durationUs);
|
|
|
|
if (durationUs) {
|
|
|
|
ReentrantMonitorAutoEnter mon(mDecoder->GetReentrantMonitor());
|
|
|
|
mDecoder->GetStateMachine()->SetDuration(durationUs);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (mOmxDecoder->HasVideo()) {
|
|
|
|
int32_t width, height;
|
|
|
|
mOmxDecoder->GetVideoParameters(&width, &height);
|
|
|
|
nsIntRect pictureRect(0, 0, width, height);
|
|
|
|
|
|
|
|
// Validate the container-reported frame and pictureRect sizes. This ensures
|
|
|
|
// that our video frame creation code doesn't overflow.
|
|
|
|
nsIntSize displaySize(width, height);
|
|
|
|
nsIntSize frameSize(width, height);
|
|
|
|
if (!nsVideoInfo::ValidateVideoRegion(frameSize, pictureRect, displaySize)) {
|
|
|
|
return NS_ERROR_FAILURE;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Video track's frame sizes will not overflow. Activate the video track.
|
|
|
|
mHasVideo = mInfo.mHasVideo = true;
|
|
|
|
mInfo.mDisplay = displaySize;
|
|
|
|
mPicture = pictureRect;
|
|
|
|
mInitialFrame = frameSize;
|
|
|
|
VideoFrameContainer* container = mDecoder->GetVideoFrameContainer();
|
|
|
|
if (container) {
|
|
|
|
container->SetCurrentFrame(gfxIntSize(displaySize.width, displaySize.height),
|
|
|
|
nullptr,
|
|
|
|
mozilla::TimeStamp::Now());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (mOmxDecoder->HasAudio()) {
|
|
|
|
int32_t numChannels, sampleRate;
|
|
|
|
mOmxDecoder->GetAudioParameters(&numChannels, &sampleRate);
|
|
|
|
mHasAudio = mInfo.mHasAudio = true;
|
|
|
|
mInfo.mAudioChannels = numChannels;
|
|
|
|
mInfo.mAudioRate = sampleRate;
|
|
|
|
}
|
|
|
|
|
|
|
|
*aInfo = mInfo;
|
|
|
|
|
|
|
|
return NS_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Resets all state related to decoding, emptying all buffers etc.
|
|
|
|
nsresult nsMediaOmxReader::ResetDecode()
|
|
|
|
{
|
2012-09-26 21:33:43 -07:00
|
|
|
nsBuiltinDecoderReader::ResetDecode();
|
|
|
|
|
|
|
|
VideoFrameContainer* container = mDecoder->GetVideoFrameContainer();
|
|
|
|
if (container) {
|
|
|
|
container->ClearCurrentFrame();
|
|
|
|
}
|
|
|
|
|
2012-09-26 21:33:43 -07:00
|
|
|
if (mLastVideoFrame) {
|
|
|
|
delete mLastVideoFrame;
|
|
|
|
mLastVideoFrame = nullptr;
|
|
|
|
}
|
|
|
|
if (mOmxDecoder) {
|
|
|
|
delete mOmxDecoder;
|
|
|
|
mOmxDecoder = nullptr;
|
|
|
|
}
|
|
|
|
return NS_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool nsMediaOmxReader::DecodeVideoFrame(bool &aKeyframeSkip,
|
|
|
|
int64_t aTimeThreshold)
|
|
|
|
{
|
|
|
|
// Record number of frames decoded and parsed. Automatically update the
|
|
|
|
// stats counters using the AutoNotifyDecoded stack-based class.
|
|
|
|
uint32_t parsed = 0, decoded = 0;
|
|
|
|
nsMediaDecoder::AutoNotifyDecoded autoNotify(mDecoder, parsed, decoded);
|
|
|
|
|
|
|
|
// Throw away the currently buffered frame if we are seeking.
|
|
|
|
if (mLastVideoFrame && mVideoSeekTimeUs != -1) {
|
|
|
|
delete mLastVideoFrame;
|
|
|
|
mLastVideoFrame = nullptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool doSeek = mVideoSeekTimeUs != -1;
|
|
|
|
if (doSeek) {
|
|
|
|
aTimeThreshold = mVideoSeekTimeUs;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Read next frame
|
|
|
|
while (true) {
|
|
|
|
MPAPI::VideoFrame frame;
|
|
|
|
frame.mGraphicBuffer = nullptr;
|
|
|
|
if (!mOmxDecoder->ReadVideo(&frame, aTimeThreshold, aKeyframeSkip, doSeek)) {
|
|
|
|
// We reached the end of the video stream. If we have a buffered
|
|
|
|
// video frame, push it the video queue using the total duration
|
|
|
|
// of the video as the end time.
|
|
|
|
if (mLastVideoFrame) {
|
|
|
|
int64_t durationUs;
|
|
|
|
mOmxDecoder->GetDuration(&durationUs);
|
|
|
|
mLastVideoFrame->mEndTime = (durationUs > mLastVideoFrame->mTime)
|
|
|
|
? durationUs
|
|
|
|
: mLastVideoFrame->mTime;
|
|
|
|
mVideoQueue.Push(mLastVideoFrame);
|
|
|
|
mLastVideoFrame = nullptr;
|
|
|
|
}
|
|
|
|
mVideoQueue.Finish();
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
mVideoSeekTimeUs = -1;
|
|
|
|
doSeek = aKeyframeSkip = false;
|
|
|
|
|
|
|
|
nsIntRect picture = mPicture;
|
|
|
|
if (frame.Y.mWidth != mInitialFrame.width ||
|
|
|
|
frame.Y.mHeight != mInitialFrame.height) {
|
|
|
|
|
|
|
|
// Frame size is different from what the container reports. This is legal,
|
|
|
|
// and we will preserve the ratio of the crop rectangle as it
|
|
|
|
// was reported relative to the picture size reported by the container.
|
|
|
|
picture.x = (mPicture.x * frame.Y.mWidth) / mInitialFrame.width;
|
|
|
|
picture.y = (mPicture.y * frame.Y.mHeight) / mInitialFrame.height;
|
|
|
|
picture.width = (frame.Y.mWidth * mPicture.width) / mInitialFrame.width;
|
|
|
|
picture.height = (frame.Y.mHeight * mPicture.height) / mInitialFrame.height;
|
|
|
|
}
|
|
|
|
|
|
|
|
// This is the approximate byte position in the stream.
|
|
|
|
int64_t pos = mDecoder->GetResource()->Tell();
|
|
|
|
|
|
|
|
VideoData *v;
|
|
|
|
if (!frame.mGraphicBuffer) {
|
|
|
|
|
|
|
|
VideoData::YCbCrBuffer b;
|
|
|
|
b.mPlanes[0].mData = static_cast<uint8_t *>(frame.Y.mData);
|
|
|
|
b.mPlanes[0].mStride = frame.Y.mStride;
|
|
|
|
b.mPlanes[0].mHeight = frame.Y.mHeight;
|
|
|
|
b.mPlanes[0].mWidth = frame.Y.mWidth;
|
|
|
|
b.mPlanes[0].mOffset = frame.Y.mOffset;
|
|
|
|
b.mPlanes[0].mSkip = frame.Y.mSkip;
|
|
|
|
|
|
|
|
b.mPlanes[1].mData = static_cast<uint8_t *>(frame.Cb.mData);
|
|
|
|
b.mPlanes[1].mStride = frame.Cb.mStride;
|
|
|
|
b.mPlanes[1].mHeight = frame.Cb.mHeight;
|
|
|
|
b.mPlanes[1].mWidth = frame.Cb.mWidth;
|
|
|
|
b.mPlanes[1].mOffset = frame.Cb.mOffset;
|
|
|
|
b.mPlanes[1].mSkip = frame.Cb.mSkip;
|
|
|
|
|
|
|
|
b.mPlanes[2].mData = static_cast<uint8_t *>(frame.Cr.mData);
|
|
|
|
b.mPlanes[2].mStride = frame.Cr.mStride;
|
|
|
|
b.mPlanes[2].mHeight = frame.Cr.mHeight;
|
|
|
|
b.mPlanes[2].mWidth = frame.Cr.mWidth;
|
|
|
|
b.mPlanes[2].mOffset = frame.Cr.mOffset;
|
|
|
|
b.mPlanes[2].mSkip = frame.Cr.mSkip;
|
|
|
|
|
|
|
|
v = VideoData::Create(mInfo,
|
|
|
|
mDecoder->GetImageContainer(),
|
|
|
|
pos,
|
|
|
|
frame.mTimeUs,
|
|
|
|
frame.mTimeUs+1, // We don't know the end time.
|
|
|
|
b,
|
|
|
|
frame.mKeyFrame,
|
|
|
|
-1,
|
|
|
|
picture);
|
|
|
|
} else {
|
|
|
|
v = VideoData::Create(mInfo,
|
|
|
|
mDecoder->GetImageContainer(),
|
|
|
|
pos,
|
|
|
|
frame.mTimeUs,
|
|
|
|
frame.mTimeUs+1, // We don't know the end time.
|
|
|
|
frame.mGraphicBuffer,
|
|
|
|
frame.mKeyFrame,
|
|
|
|
-1,
|
|
|
|
picture);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!v) {
|
|
|
|
NS_WARNING("Unable to create VideoData");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
parsed++;
|
|
|
|
decoded++;
|
|
|
|
NS_ASSERTION(decoded <= parsed, "Expect to decode fewer frames than parsed in MediaPlugin...");
|
|
|
|
|
|
|
|
// Seeking hack
|
|
|
|
if (mLastVideoFrame && mLastVideoFrame->mTime > v->mTime) {
|
|
|
|
delete mLastVideoFrame;
|
|
|
|
mLastVideoFrame = v;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Since MPAPI doesn't give us the end time of frames, we keep one frame
|
|
|
|
// buffered in nsMediaOmxReader and push it into the queue as soon
|
|
|
|
// we read the following frame so we can use that frame's start time as
|
|
|
|
// the end time of the buffered frame.
|
|
|
|
if (!mLastVideoFrame) {
|
|
|
|
mLastVideoFrame = v;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
mLastVideoFrame->mEndTime = v->mTime;
|
|
|
|
|
|
|
|
mVideoQueue.Push(mLastVideoFrame);
|
|
|
|
|
|
|
|
// Buffer the current frame we just decoded.
|
|
|
|
mLastVideoFrame = v;
|
|
|
|
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool nsMediaOmxReader::DecodeAudioData()
|
|
|
|
{
|
|
|
|
NS_ASSERTION(mDecoder->OnDecodeThread(), "Should be on decode thread.");
|
|
|
|
|
|
|
|
// This is the approximate byte position in the stream.
|
|
|
|
int64_t pos = mDecoder->GetResource()->Tell();
|
|
|
|
|
|
|
|
// Read next frame
|
|
|
|
MPAPI::AudioFrame frame;
|
|
|
|
if (!mOmxDecoder->ReadAudio(&frame, mAudioSeekTimeUs)) {
|
|
|
|
mAudioQueue.Finish();
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
mAudioSeekTimeUs = -1;
|
|
|
|
|
|
|
|
// Ignore empty buffer which stagefright media read will sporadically return
|
|
|
|
if (frame.mSize == 0) {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
nsAutoArrayPtr<AudioDataValue> buffer(new AudioDataValue[frame.mSize/2] );
|
|
|
|
memcpy(buffer.get(), frame.mData, frame.mSize);
|
|
|
|
|
|
|
|
uint32_t frames = frame.mSize / (2 * frame.mAudioChannels);
|
|
|
|
CheckedInt64 duration = FramesToUsecs(frames, frame.mAudioSampleRate);
|
|
|
|
if (!duration.isValid()) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
mAudioQueue.Push(new AudioData(pos,
|
|
|
|
frame.mTimeUs,
|
|
|
|
duration.value(),
|
|
|
|
frames,
|
|
|
|
buffer.forget(),
|
|
|
|
frame.mAudioChannels));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
nsresult nsMediaOmxReader::Seek(int64_t aTarget, int64_t aStartTime, int64_t aEndTime, int64_t aCurrentTime)
|
|
|
|
{
|
|
|
|
NS_ASSERTION(mDecoder->OnDecodeThread(), "Should be on decode thread.");
|
|
|
|
|
|
|
|
mVideoQueue.Reset();
|
|
|
|
mAudioQueue.Reset();
|
|
|
|
|
|
|
|
mAudioSeekTimeUs = mVideoSeekTimeUs = aTarget;
|
|
|
|
|
|
|
|
return DecodeToTarget(aTarget);
|
|
|
|
}
|
|
|
|
|
|
|
|
static uint64_t BytesToTime(int64_t offset, uint64_t length, uint64_t durationUs) {
|
|
|
|
double perc = double(offset) / double(length);
|
|
|
|
if (perc > 1.0)
|
|
|
|
perc = 1.0;
|
|
|
|
return uint64_t(double(durationUs) * perc);
|
|
|
|
}
|
|
|
|
|
|
|
|
nsresult nsMediaOmxReader::GetBuffered(nsTimeRanges* aBuffered, int64_t aStartTime)
|
|
|
|
{
|
|
|
|
if (!mOmxDecoder)
|
|
|
|
return NS_OK;
|
|
|
|
|
|
|
|
MediaResource* stream = mOmxDecoder->GetResource();
|
|
|
|
|
|
|
|
int64_t durationUs = 0;
|
|
|
|
mOmxDecoder->GetDuration(&durationUs);
|
|
|
|
|
|
|
|
// Nothing to cache if the media takes 0us to play.
|
|
|
|
if (!durationUs)
|
|
|
|
return NS_OK;
|
|
|
|
|
|
|
|
// Special case completely cached files. This also handles local files.
|
|
|
|
if (stream->IsDataCachedToEndOfResource(0)) {
|
|
|
|
aBuffered->Add(0, durationUs);
|
|
|
|
return NS_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
int64_t totalBytes = stream->GetLength();
|
|
|
|
|
|
|
|
// If we can't determine the total size, pretend that we have nothing
|
|
|
|
// buffered. This will put us in a state of eternally-low-on-undecoded-data
|
|
|
|
// which is not get, but about the best we can do.
|
|
|
|
if (totalBytes == -1)
|
|
|
|
return NS_OK;
|
|
|
|
|
|
|
|
int64_t startOffset = stream->GetNextCachedData(0);
|
|
|
|
while (startOffset >= 0) {
|
|
|
|
int64_t endOffset = stream->GetCachedDataEnd(startOffset);
|
|
|
|
// Bytes [startOffset..endOffset] are cached.
|
|
|
|
NS_ASSERTION(startOffset >= 0, "Integer underflow in GetBuffered");
|
|
|
|
NS_ASSERTION(endOffset >= 0, "Integer underflow in GetBuffered");
|
|
|
|
|
|
|
|
uint64_t startUs = BytesToTime(startOffset, totalBytes, durationUs);
|
|
|
|
uint64_t endUs = BytesToTime(endOffset, totalBytes, durationUs);
|
|
|
|
if (startUs != endUs) {
|
|
|
|
aBuffered->Add((double)startUs / USECS_PER_S, (double)endUs / USECS_PER_S);
|
|
|
|
}
|
|
|
|
startOffset = stream->GetNextCachedData(endOffset);
|
|
|
|
}
|
|
|
|
return NS_OK;
|
|
|
|
}
|