/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */ /* vim:set ts=2 sw=2 sts=2 et cindent: */ /* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this file, * You can obtain one at http://mozilla.org/MPL/2.0/. */ #include "MediaOmxReader.h" #include "MediaDecoderStateMachine.h" #include "mozilla/TimeStamp.h" #include "mozilla/dom/TimeRanges.h" #include "MediaResource.h" #include "VideoUtils.h" #include "MediaOmxDecoder.h" #include "AbstractMediaDecoder.h" #include "AudioChannelService.h" #include "OmxDecoder.h" #include "MPAPI.h" #include "gfx2DGlue.h" #include "MediaStreamSource.h" #define MAX_DROPPED_FRAMES 25 // Try not to spend more than this much time in a single call to DecodeVideoFrame. #define MAX_VIDEO_DECODE_SECONDS 0.1 using namespace mozilla::gfx; using namespace android; namespace mozilla { #ifdef PR_LOGGING extern PRLogModuleInfo* gMediaDecoderLog; #define DECODER_LOG(type, msg) PR_LOG(gMediaDecoderLog, type, msg) #else #define DECODER_LOG(type, msg) #endif class OmxReaderProcessCachedDataTask : public Task { public: OmxReaderProcessCachedDataTask(MediaOmxReader* aOmxReader, int64_t aOffset) : mOmxReader(aOmxReader), mOffset(aOffset) { } void Run() { MOZ_ASSERT(!NS_IsMainThread()); MOZ_ASSERT(mOmxReader.get()); mOmxReader->ProcessCachedData(mOffset, false); } private: nsRefPtr mOmxReader; int64_t mOffset; }; // When loading an MP3 stream from a file, we need to parse the file's // content to find its duration. Reading files of 100 MiB or more can // delay the player app noticably, so the file is read and decoded in // smaller chunks. // // We first read on the decode thread, but parsing must be done on the // main thread. After we read the file's initial MiBs in the decode // thread, an instance of this class is scheduled to the main thread for // parsing the MP3 stream. The decode thread waits until it has finished. // // If there is more data available from the file, the runnable dispatches // a task to the IO thread for retrieving the next chunk of data, and // the IO task dispatches a runnable to the main thread for parsing the // data. This goes on until all of the MP3 file has been parsed. class OmxReaderNotifyDataArrivedRunnable : public nsRunnable { public: OmxReaderNotifyDataArrivedRunnable(MediaOmxReader* aOmxReader, const char* aBuffer, uint64_t aLength, int64_t aOffset, uint64_t aFullLength) : mOmxReader(aOmxReader), mBuffer(aBuffer), mLength(aLength), mOffset(aOffset), mFullLength(aFullLength) { MOZ_ASSERT(mOmxReader.get()); MOZ_ASSERT(mBuffer.get() || !mLength); } NS_IMETHOD Run() { NS_ASSERTION(NS_IsMainThread(), "Should be on main thread."); NotifyDataArrived(); return NS_OK; } private: void NotifyDataArrived() { if (mOmxReader->IsShutdown()) { return; } const char* buffer = mBuffer.get(); while (mLength) { uint32_t length = std::min(mLength, UINT32_MAX); mOmxReader->NotifyDataArrived(buffer, length, mOffset); buffer += length; mLength -= length; mOffset += length; } if (mOffset < mFullLength) { // We cannot read data in the main thread because it // might block for too long. Instead we post an IO task // to the IO thread if there is more data available. XRE_GetIOMessageLoop()->PostTask(FROM_HERE, new OmxReaderProcessCachedDataTask(mOmxReader.get(), mOffset)); } } nsRefPtr mOmxReader; nsAutoArrayPtr mBuffer; uint64_t mLength; int64_t mOffset; uint64_t mFullLength; }; void MediaOmxReader::CancelProcessCachedData() { MOZ_ASSERT(NS_IsMainThread()); MutexAutoLock lock(mMutex); mIsShutdown = true; } MediaOmxReader::MediaOmxReader(AbstractMediaDecoder *aDecoder) : MediaOmxCommonReader(aDecoder) , mMutex("MediaOmxReader.Data") , mMP3FrameParser(-1) , mHasVideo(false) , mHasAudio(false) , mVideoSeekTimeUs(-1) , mAudioSeekTimeUs(-1) , mSkipCount(0) , mUseParserDuration(false) , mLastParserDuration(-1) , mIsShutdown(false) , mIsWaitingResources(false) { #ifdef PR_LOGGING if (!gMediaDecoderLog) { gMediaDecoderLog = PR_NewLogModule("MediaDecoder"); } #endif mAudioChannel = dom::AudioChannelService::GetDefaultAudioChannel(); } MediaOmxReader::~MediaOmxReader() { } nsresult MediaOmxReader::Init(MediaDecoderReader* aCloneDonor) { return NS_OK; } void MediaOmxReader::ReleaseDecoder() { if (mOmxDecoder.get()) { mOmxDecoder->ReleaseDecoder(); } mOmxDecoder.clear(); } nsRefPtr MediaOmxReader::Shutdown() { nsCOMPtr cancelEvent = NS_NewRunnableMethod(this, &MediaOmxReader::CancelProcessCachedData); NS_DispatchToMainThread(cancelEvent); nsRefPtr p = MediaDecoderReader::Shutdown(); // Wait for the superclass to finish tearing things down before releasing // the decoder on the main thread. nsCOMPtr mt; NS_GetMainThread(getter_AddRefs(mt)); p->Then(mt.get(), __func__, this, &MediaOmxReader::ReleaseDecoder, &MediaOmxReader::ReleaseDecoder); return p; } bool MediaOmxReader::IsWaitingMediaResources() { return mIsWaitingResources; } void MediaOmxReader::UpdateIsWaitingMediaResources() { if (mOmxDecoder.get()) { mIsWaitingResources = mOmxDecoder->IsWaitingMediaResources(); } else { mIsWaitingResources = false; } } bool MediaOmxReader::IsDormantNeeded() { if (!mOmxDecoder.get()) { return false; } return mOmxDecoder->IsDormantNeeded(); } void MediaOmxReader::ReleaseMediaResources() { ResetDecode(); // Before freeing a video codec, all video buffers needed to be released // even from graphics pipeline. VideoFrameContainer* container = mDecoder->GetVideoFrameContainer(); if (container) { container->ClearCurrentFrame(); } if (mOmxDecoder.get()) { mOmxDecoder->ReleaseMediaResources(); } } nsresult MediaOmxReader::InitOmxDecoder() { if (!mOmxDecoder.get()) { //register sniffers, if they are not registered in this process. DataSource::RegisterDefaultSniffers(); mDecoder->GetResource()->SetReadMode(MediaCacheStream::MODE_METADATA); sp dataSource = new MediaStreamSource(mDecoder->GetResource()); dataSource->initCheck(); mExtractor = MediaExtractor::Create(dataSource); if (!mExtractor.get()) { return NS_ERROR_FAILURE; } mOmxDecoder = new OmxDecoder(mDecoder->GetResource(), mDecoder); if (!mOmxDecoder->Init(mExtractor)) { return NS_ERROR_FAILURE; } } return NS_OK; } void MediaOmxReader::PreReadMetadata() { UpdateIsWaitingMediaResources(); } nsresult MediaOmxReader::ReadMetadata(MediaInfo* aInfo, MetadataTags** aTags) { NS_ASSERTION(mDecoder->OnDecodeThread(), "Should be on decode thread."); EnsureActive(); *aTags = nullptr; // Initialize the internal OMX Decoder. nsresult rv = InitOmxDecoder(); if (NS_FAILED(rv)) { return rv; } bool isMP3 = mDecoder->GetResource()->GetContentType().EqualsASCII(AUDIO_MP3); if (isMP3) { // When read sdcard's file on b2g platform at constructor, // the mDecoder->GetResource()->GetLength() would return -1. // Delay set the total duration on this function. mMP3FrameParser.SetLength(mDecoder->GetResource()->GetLength()); ProcessCachedData(0, true); } if (!mOmxDecoder->AllocateMediaResources()) { return NS_ERROR_FAILURE; } // Bug 1050667, both MediaDecoderStateMachine and MediaOmxReader // relies on IsWaitingMediaResources() function. And the waiting state will be // changed by binder thread, so we store the waiting state in a cache value to // make them in consistent state. UpdateIsWaitingMediaResources(); if (IsWaitingMediaResources()) { return NS_OK; } // After resources are available, set the metadata. if (!mOmxDecoder->EnsureMetadata()) { return NS_ERROR_FAILURE; } if (isMP3 && mMP3FrameParser.IsMP3()) { int64_t duration = mMP3FrameParser.GetDuration(); // The MP3FrameParser may reported a duration; // return -1 if no frame has been parsed. if (duration >= 0) { ReentrantMonitorAutoEnter mon(mDecoder->GetReentrantMonitor()); mUseParserDuration = true; mLastParserDuration = duration; mDecoder->SetMediaDuration(mLastParserDuration); } } else { // Set the total duration (the max of the audio and video track). int64_t durationUs; mOmxDecoder->GetDuration(&durationUs); if (durationUs) { ReentrantMonitorAutoEnter mon(mDecoder->GetReentrantMonitor()); mDecoder->SetMediaDuration(durationUs); } } if (mOmxDecoder->HasVideo()) { int32_t displayWidth, displayHeight, width, height; mOmxDecoder->GetVideoParameters(&displayWidth, &displayHeight, &width, &height); nsIntRect pictureRect(0, 0, width, height); // Validate the container-reported frame and pictureRect sizes. This ensures // that our video frame creation code doesn't overflow. nsIntSize displaySize(displayWidth, displayHeight); nsIntSize frameSize(width, height); if (!IsValidVideoRegion(frameSize, pictureRect, displaySize)) { return NS_ERROR_FAILURE; } // Video track's frame sizes will not overflow. Activate the video track. mHasVideo = mInfo.mVideo.mHasVideo = true; mInfo.mVideo.mDisplay = displaySize; mPicture = pictureRect; mInitialFrame = frameSize; VideoFrameContainer* container = mDecoder->GetVideoFrameContainer(); if (container) { container->SetCurrentFrame(gfxIntSize(displaySize.width, displaySize.height), nullptr, mozilla::TimeStamp::Now()); } } if (mOmxDecoder->HasAudio()) { int32_t numChannels, sampleRate; mOmxDecoder->GetAudioParameters(&numChannels, &sampleRate); mHasAudio = mInfo.mAudio.mHasAudio = true; mInfo.mAudio.mChannels = numChannels; mInfo.mAudio.mRate = sampleRate; } *aInfo = mInfo; #ifdef MOZ_AUDIO_OFFLOAD CheckAudioOffload(); #endif return NS_OK; } bool MediaOmxReader::IsMediaSeekable() { // Check the MediaExtract flag if the source is seekable. return (mExtractor->flags() & MediaExtractor::CAN_SEEK); } bool MediaOmxReader::DecodeVideoFrame(bool &aKeyframeSkip, int64_t aTimeThreshold) { NS_ASSERTION(mDecoder->OnDecodeThread(), "Should be on decode thread."); EnsureActive(); // Record number of frames decoded and parsed. Automatically update the // stats counters using the AutoNotifyDecoded stack-based class. uint32_t parsed = 0, decoded = 0; AbstractMediaDecoder::AutoNotifyDecoded autoNotify(mDecoder, parsed, decoded); bool doSeek = mVideoSeekTimeUs != -1; if (doSeek) { aTimeThreshold = mVideoSeekTimeUs; } TimeStamp start = TimeStamp::Now(); // Read next frame. Don't let this loop run for too long. while ((TimeStamp::Now() - start) < TimeDuration::FromSeconds(MAX_VIDEO_DECODE_SECONDS)) { MPAPI::VideoFrame frame; frame.mGraphicBuffer = nullptr; frame.mShouldSkip = false; if (!mOmxDecoder->ReadVideo(&frame, aTimeThreshold, aKeyframeSkip, doSeek)) { return false; } doSeek = false; mVideoSeekTimeUs = -1; // Ignore empty buffer which stagefright media read will sporadically return if (frame.mSize == 0 && !frame.mGraphicBuffer) { continue; } parsed++; if (frame.mShouldSkip && mSkipCount < MAX_DROPPED_FRAMES) { mSkipCount++; continue; } mSkipCount = 0; aKeyframeSkip = false; IntRect picture = ToIntRect(mPicture); if (frame.Y.mWidth != mInitialFrame.width || frame.Y.mHeight != mInitialFrame.height) { // Frame size is different from what the container reports. This is legal, // and we will preserve the ratio of the crop rectangle as it // was reported relative to the picture size reported by the container. picture.x = (mPicture.x * frame.Y.mWidth) / mInitialFrame.width; picture.y = (mPicture.y * frame.Y.mHeight) / mInitialFrame.height; picture.width = (frame.Y.mWidth * mPicture.width) / mInitialFrame.width; picture.height = (frame.Y.mHeight * mPicture.height) / mInitialFrame.height; } // This is the approximate byte position in the stream. int64_t pos = mDecoder->GetResource()->Tell(); nsRefPtr v; if (!frame.mGraphicBuffer) { VideoData::YCbCrBuffer b; b.mPlanes[0].mData = static_cast(frame.Y.mData); b.mPlanes[0].mStride = frame.Y.mStride; b.mPlanes[0].mHeight = frame.Y.mHeight; b.mPlanes[0].mWidth = frame.Y.mWidth; b.mPlanes[0].mOffset = frame.Y.mOffset; b.mPlanes[0].mSkip = frame.Y.mSkip; b.mPlanes[1].mData = static_cast(frame.Cb.mData); b.mPlanes[1].mStride = frame.Cb.mStride; b.mPlanes[1].mHeight = frame.Cb.mHeight; b.mPlanes[1].mWidth = frame.Cb.mWidth; b.mPlanes[1].mOffset = frame.Cb.mOffset; b.mPlanes[1].mSkip = frame.Cb.mSkip; b.mPlanes[2].mData = static_cast(frame.Cr.mData); b.mPlanes[2].mStride = frame.Cr.mStride; b.mPlanes[2].mHeight = frame.Cr.mHeight; b.mPlanes[2].mWidth = frame.Cr.mWidth; b.mPlanes[2].mOffset = frame.Cr.mOffset; b.mPlanes[2].mSkip = frame.Cr.mSkip; v = VideoData::Create(mInfo.mVideo, mDecoder->GetImageContainer(), pos, frame.mTimeUs, 1, // We don't know the duration. b, frame.mKeyFrame, -1, picture); } else { v = VideoData::Create(mInfo.mVideo, mDecoder->GetImageContainer(), pos, frame.mTimeUs, 1, // We don't know the duration. frame.mGraphicBuffer, frame.mKeyFrame, -1, picture); } if (!v) { NS_WARNING("Unable to create VideoData"); return false; } decoded++; NS_ASSERTION(decoded <= parsed, "Expect to decode fewer frames than parsed in OMX decoder..."); mVideoQueue.Push(v); break; } return true; } void MediaOmxReader::NotifyDataArrived(const char* aBuffer, uint32_t aLength, int64_t aOffset) { MOZ_ASSERT(NS_IsMainThread()); if (IsShutdown()) { return; } if (HasVideo()) { return; } if (!mMP3FrameParser.NeedsData()) { return; } mMP3FrameParser.Parse(aBuffer, aLength, aOffset); int64_t duration = mMP3FrameParser.GetDuration(); ReentrantMonitorAutoEnter mon(mDecoder->GetReentrantMonitor()); if (duration != mLastParserDuration && mUseParserDuration) { mLastParserDuration = duration; mDecoder->UpdateEstimatedMediaDuration(mLastParserDuration); } } bool MediaOmxReader::DecodeAudioData() { NS_ASSERTION(mDecoder->OnDecodeThread(), "Should be on decode thread."); EnsureActive(); // This is the approximate byte position in the stream. int64_t pos = mDecoder->GetResource()->Tell(); // Read next frame MPAPI::AudioFrame source; if (!mOmxDecoder->ReadAudio(&source, mAudioSeekTimeUs)) { return false; } mAudioSeekTimeUs = -1; // Ignore empty buffer which stagefright media read will sporadically return if (source.mSize == 0) { return true; } uint32_t frames = source.mSize / (source.mAudioChannels * sizeof(AudioDataValue)); typedef AudioCompactor::NativeCopy OmxCopy; return mAudioCompactor.Push(pos, source.mTimeUs, source.mAudioSampleRate, frames, source.mAudioChannels, OmxCopy(static_cast(source.mData), source.mSize, source.mAudioChannels)); } nsRefPtr MediaOmxReader::Seek(int64_t aTarget, int64_t aStartTime, int64_t aEndTime, int64_t aCurrentTime) { NS_ASSERTION(mDecoder->OnDecodeThread(), "Should be on decode thread."); EnsureActive(); VideoFrameContainer* container = mDecoder->GetVideoFrameContainer(); if (container && container->GetImageContainer()) { container->GetImageContainer()->ClearAllImagesExceptFront(); } if (mHasAudio && mHasVideo) { // The OMXDecoder seeks/demuxes audio and video streams separately. So if // we seek both audio and video to aTarget, the audio stream can typically // seek closer to the seek target, since typically every audio block is // a sync point, whereas for video there are only keyframes once every few // seconds. So if we have both audio and video, we must seek the video // stream to the preceeding keyframe first, get the stream time, and then // seek the audio stream to match the video stream's time. Otherwise, the // audio and video streams won't be in sync after the seek. mVideoSeekTimeUs = aTarget; const VideoData* v = DecodeToFirstVideoData(); mAudioSeekTimeUs = v ? v->mTime : aTarget; } else { mAudioSeekTimeUs = mVideoSeekTimeUs = aTarget; } return SeekPromise::CreateAndResolve(true, __func__); } void MediaOmxReader::SetIdle() { if (!mOmxDecoder.get()) { return; } mOmxDecoder->Pause(); } void MediaOmxReader::EnsureActive() { if (!mOmxDecoder.get()) { return; } DebugOnly result = mOmxDecoder->Play(); NS_ASSERTION(result == NS_OK, "OmxDecoder should be in play state to continue decoding"); } int64_t MediaOmxReader::ProcessCachedData(int64_t aOffset, bool aWaitForCompletion) { // Could run on decoder thread or IO thread. if (IsShutdown()) { return -1; } // We read data in chunks of 32 KiB. We can reduce this // value if media, such as sdcards, is too slow. // Because of SD card's slowness, need to keep sReadSize to small size. // See Bug 914870. static const int64_t sReadSize = 32 * 1024; NS_ASSERTION(!NS_IsMainThread(), "Should not be on main thread."); MOZ_ASSERT(mDecoder->GetResource()); int64_t resourceLength = mDecoder->GetResource()->GetCachedDataEnd(0); NS_ENSURE_TRUE(resourceLength >= 0, -1); if (aOffset >= resourceLength) { return 0; // Cache is empty, nothing to do } int64_t bufferLength = std::min(resourceLength-aOffset, sReadSize); nsAutoArrayPtr buffer(new char[bufferLength]); nsresult rv = mDecoder->GetResource()->ReadFromCache(buffer.get(), aOffset, bufferLength); NS_ENSURE_SUCCESS(rv, -1); nsRefPtr runnable( new OmxReaderNotifyDataArrivedRunnable(this, buffer.forget(), bufferLength, aOffset, resourceLength)); if (aWaitForCompletion) { rv = NS_DispatchToMainThread(runnable.get(), NS_DISPATCH_SYNC); } else { rv = NS_DispatchToMainThread(runnable.get()); } NS_ENSURE_SUCCESS(rv, -1); return resourceLength - aOffset - bufferLength; } android::sp MediaOmxReader::GetAudioOffloadTrack() { if (!mOmxDecoder.get()) { return nullptr; } return mOmxDecoder->GetAudioOffloadTrack(); } } // namespace mozilla