mirror of
https://gitlab.winehq.org/wine/wine-gecko.git
synced 2024-09-13 09:24:08 -07:00
373 lines
11 KiB
C++
373 lines
11 KiB
C++
/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
|
|
/* vim: set ts=8 sts=2 et sw=2 tw=80: */
|
|
/* This Source Code Form is subject to the terms of the Mozilla Public
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
|
|
|
|
#include "ContainerParser.h"
|
|
|
|
#include "WebMBufferedParser.h"
|
|
#include "mozilla/Endian.h"
|
|
#include "mp4_demuxer/MoofParser.h"
|
|
#include "prlog.h"
|
|
#include "MediaData.h"
|
|
#ifdef MOZ_FMP4
|
|
#include "MP4Stream.h"
|
|
#endif
|
|
#include "SourceBufferResource.h"
|
|
|
|
#ifdef PR_LOGGING
|
|
extern PRLogModuleInfo* GetMediaSourceLog();
|
|
|
|
/* Polyfill __func__ on MSVC to pass to the log. */
|
|
#ifdef _MSC_VER
|
|
#define __func__ __FUNCTION__
|
|
#endif
|
|
|
|
#define STRINGIFY(x) #x
|
|
#define TOSTRING(x) STRINGIFY(x)
|
|
#define MSE_DEBUG(name, arg, ...) PR_LOG(GetMediaSourceLog(), PR_LOG_DEBUG, (TOSTRING(name) "(%p:%s)::%s: " arg, this, mType.get(), __func__, ##__VA_ARGS__))
|
|
#else
|
|
#define MSE_DEBUG(...)
|
|
#endif
|
|
|
|
namespace mozilla {
|
|
|
|
ContainerParser::ContainerParser(const nsACString& aType)
|
|
: mHasInitData(false)
|
|
, mType(aType)
|
|
{
|
|
}
|
|
|
|
bool
|
|
ContainerParser::IsInitSegmentPresent(LargeDataBuffer* aData)
|
|
{
|
|
MSE_DEBUG(ContainerParser, "aLength=%u [%x%x%x%x]",
|
|
aData->Length(),
|
|
aData->Length() > 0 ? (*aData)[0] : 0,
|
|
aData->Length() > 1 ? (*aData)[1] : 0,
|
|
aData->Length() > 2 ? (*aData)[2] : 0,
|
|
aData->Length() > 3 ? (*aData)[3] : 0);
|
|
return false;
|
|
}
|
|
|
|
bool
|
|
ContainerParser::IsMediaSegmentPresent(LargeDataBuffer* aData)
|
|
{
|
|
MSE_DEBUG(ContainerParser, "aLength=%u [%x%x%x%x]",
|
|
aData->Length(),
|
|
aData->Length() > 0 ? (*aData)[0] : 0,
|
|
aData->Length() > 1 ? (*aData)[1] : 0,
|
|
aData->Length() > 2 ? (*aData)[2] : 0,
|
|
aData->Length() > 3 ? (*aData)[3] : 0);
|
|
return false;
|
|
}
|
|
|
|
bool
|
|
ContainerParser::ParseStartAndEndTimestamps(LargeDataBuffer* aData,
|
|
int64_t& aStart, int64_t& aEnd)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
bool
|
|
ContainerParser::TimestampsFuzzyEqual(int64_t aLhs, int64_t aRhs)
|
|
{
|
|
return llabs(aLhs - aRhs) <= GetRoundingError();
|
|
}
|
|
|
|
int64_t
|
|
ContainerParser::GetRoundingError()
|
|
{
|
|
NS_WARNING("Using default ContainerParser::GetRoundingError implementation");
|
|
return 0;
|
|
}
|
|
|
|
bool
|
|
ContainerParser::HasCompleteInitData()
|
|
{
|
|
return mHasInitData && !!mInitData->Length();
|
|
}
|
|
|
|
LargeDataBuffer*
|
|
ContainerParser::InitData()
|
|
{
|
|
return mInitData;
|
|
}
|
|
|
|
class WebMContainerParser : public ContainerParser {
|
|
public:
|
|
explicit WebMContainerParser(const nsACString& aType)
|
|
: ContainerParser(aType)
|
|
, mParser(0)
|
|
, mOffset(0)
|
|
{}
|
|
|
|
static const unsigned NS_PER_USEC = 1000;
|
|
static const unsigned USEC_PER_SEC = 1000000;
|
|
|
|
bool IsInitSegmentPresent(LargeDataBuffer* aData)
|
|
{
|
|
ContainerParser::IsInitSegmentPresent(aData);
|
|
// XXX: This is overly primitive, needs to collect data as it's appended
|
|
// to the SB and handle, rather than assuming everything is present in a
|
|
// single aData segment.
|
|
// 0x1a45dfa3 // EBML
|
|
// ...
|
|
// DocType == "webm"
|
|
// ...
|
|
// 0x18538067 // Segment (must be "unknown" size or contain a value large
|
|
// enough to include the Segment Information and Tracks
|
|
// elements that follow)
|
|
// 0x1549a966 // -> Segment Info
|
|
// 0x1654ae6b // -> One or more Tracks
|
|
if (aData->Length() >= 4 &&
|
|
(*aData)[0] == 0x1a && (*aData)[1] == 0x45 && (*aData)[2] == 0xdf &&
|
|
(*aData)[3] == 0xa3) {
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool IsMediaSegmentPresent(LargeDataBuffer* aData)
|
|
{
|
|
ContainerParser::IsMediaSegmentPresent(aData);
|
|
// XXX: This is overly primitive, needs to collect data as it's appended
|
|
// to the SB and handle, rather than assuming everything is present in a
|
|
// single aData segment.
|
|
// 0x1a45dfa3 // EBML
|
|
// ...
|
|
// DocType == "webm"
|
|
// ...
|
|
// 0x18538067 // Segment (must be "unknown" size)
|
|
// 0x1549a966 // -> Segment Info
|
|
// 0x1654ae6b // -> One or more Tracks
|
|
if (aData->Length() >= 4 &&
|
|
(*aData)[0] == 0x1f && (*aData)[1] == 0x43 && (*aData)[2] == 0xb6 &&
|
|
(*aData)[3] == 0x75) {
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool ParseStartAndEndTimestamps(LargeDataBuffer* aData,
|
|
int64_t& aStart, int64_t& aEnd)
|
|
{
|
|
bool initSegment = IsInitSegmentPresent(aData);
|
|
if (initSegment) {
|
|
mOffset = 0;
|
|
mParser = WebMBufferedParser(0);
|
|
mOverlappedMapping.Clear();
|
|
mInitData = new LargeDataBuffer();
|
|
mResource = new SourceBufferResource(NS_LITERAL_CSTRING("video/webm"));
|
|
}
|
|
|
|
// XXX if it only adds new mappings, overlapped but not available
|
|
// (e.g. overlap < 0) frames are "lost" from the reported mappings here.
|
|
nsTArray<WebMTimeDataOffset> mapping;
|
|
mapping.AppendElements(mOverlappedMapping);
|
|
mOverlappedMapping.Clear();
|
|
ReentrantMonitor dummy("dummy");
|
|
mParser.Append(aData->Elements(), aData->Length(), mapping, dummy);
|
|
if (mResource) {
|
|
mResource->AppendData(aData);
|
|
}
|
|
|
|
// XXX This is a bit of a hack. Assume if there are no timecodes
|
|
// present and it's an init segment that it's _just_ an init segment.
|
|
// We should be more precise.
|
|
if (initSegment || !HasCompleteInitData()) {
|
|
if (mParser.mInitEndOffset > 0) {
|
|
MOZ_ASSERT(mParser.mInitEndOffset <= mResource->GetLength());
|
|
if (!mInitData->SetLength(mParser.mInitEndOffset)) {
|
|
// Super unlikely OOM
|
|
return false;
|
|
}
|
|
char* buffer = reinterpret_cast<char*>(mInitData->Elements());
|
|
mResource->ReadFromCache(buffer, 0, mParser.mInitEndOffset);
|
|
MSE_DEBUG(WebMContainerParser, "Stashed init of %u bytes.",
|
|
mParser.mInitEndOffset);
|
|
mResource = nullptr;
|
|
} else {
|
|
MSE_DEBUG(WebMContainerParser, "Incomplete init found.");
|
|
}
|
|
mHasInitData = true;
|
|
}
|
|
mOffset += aData->Length();
|
|
|
|
if (mapping.IsEmpty()) {
|
|
return false;
|
|
}
|
|
|
|
// Exclude frames that we don't enough data to cover the end of.
|
|
uint32_t endIdx = mapping.Length() - 1;
|
|
while (mOffset < mapping[endIdx].mEndOffset && endIdx > 0) {
|
|
endIdx -= 1;
|
|
}
|
|
|
|
if (endIdx == 0) {
|
|
return false;
|
|
}
|
|
|
|
uint64_t frameDuration = mapping[endIdx].mTimecode - mapping[endIdx - 1].mTimecode;
|
|
aStart = mapping[0].mTimecode / NS_PER_USEC;
|
|
aEnd = (mapping[endIdx].mTimecode + frameDuration) / NS_PER_USEC;
|
|
|
|
MSE_DEBUG(WebMContainerParser, "[%lld, %lld] [fso=%lld, leo=%lld, l=%u endIdx=%u]",
|
|
aStart, aEnd, mapping[0].mSyncOffset, mapping[endIdx].mEndOffset, mapping.Length(), endIdx);
|
|
|
|
mapping.RemoveElementsAt(0, endIdx + 1);
|
|
mOverlappedMapping.AppendElements(mapping);
|
|
|
|
return true;
|
|
}
|
|
|
|
int64_t GetRoundingError()
|
|
{
|
|
int64_t error = mParser.GetTimecodeScale() / NS_PER_USEC;
|
|
return error * 2;
|
|
}
|
|
|
|
private:
|
|
WebMBufferedParser mParser;
|
|
nsTArray<WebMTimeDataOffset> mOverlappedMapping;
|
|
int64_t mOffset;
|
|
};
|
|
|
|
#ifdef MOZ_FMP4
|
|
class MP4ContainerParser : public ContainerParser {
|
|
public:
|
|
explicit MP4ContainerParser(const nsACString& aType)
|
|
: ContainerParser(aType)
|
|
, mMonitor("MP4ContainerParser Index Monitor")
|
|
{}
|
|
|
|
bool IsInitSegmentPresent(LargeDataBuffer* aData)
|
|
{
|
|
ContainerParser::IsInitSegmentPresent(aData);
|
|
// Each MP4 atom has a chunk size and chunk type. The root chunk in an MP4
|
|
// file is the 'ftyp' atom followed by a file type. We just check for a
|
|
// vaguely valid 'ftyp' atom.
|
|
|
|
if (aData->Length() < 8) {
|
|
return false;
|
|
}
|
|
|
|
uint32_t chunk_size = BigEndian::readUint32(aData->Elements());
|
|
if (chunk_size < 8) {
|
|
return false;
|
|
}
|
|
|
|
return (*aData)[4] == 'f' && (*aData)[5] == 't' && (*aData)[6] == 'y' &&
|
|
(*aData)[7] == 'p';
|
|
}
|
|
|
|
bool IsMediaSegmentPresent(LargeDataBuffer* aData)
|
|
{
|
|
ContainerParser::IsMediaSegmentPresent(aData);
|
|
if (aData->Length() < 8) {
|
|
return false;
|
|
}
|
|
|
|
uint32_t chunk_size = BigEndian::readUint32(aData->Elements());
|
|
if (chunk_size < 8) {
|
|
return false;
|
|
}
|
|
|
|
return ((*aData)[4] == 'm' && (*aData)[5] == 'o' && (*aData)[6] == 'o' &&
|
|
(*aData)[7] == 'f') ||
|
|
((*aData)[4] == 's' && (*aData)[5] == 't' && (*aData)[6] == 'y' &&
|
|
(*aData)[7] == 'p');
|
|
}
|
|
|
|
bool ParseStartAndEndTimestamps(LargeDataBuffer* aData,
|
|
int64_t& aStart, int64_t& aEnd)
|
|
{
|
|
MonitorAutoLock mon(mMonitor); // We're not actually racing against anything,
|
|
// but mParser requires us to hold a monitor.
|
|
bool initSegment = IsInitSegmentPresent(aData);
|
|
if (initSegment) {
|
|
mResource = new SourceBufferResource(NS_LITERAL_CSTRING("video/mp4"));
|
|
mStream = new MP4Stream(mResource);
|
|
// We use a timestampOffset of 0 for ContainerParser, and require
|
|
// consumers of ParseStartAndEndTimestamps to add their timestamp offset
|
|
// manually. This allows the ContainerParser to be shared across different
|
|
// timestampOffsets.
|
|
mParser = new mp4_demuxer::MoofParser(mStream, 0, /* aIsAudio = */ false, &mMonitor);
|
|
mInitData = new LargeDataBuffer();
|
|
} else if (!mStream || !mParser) {
|
|
return false;
|
|
}
|
|
|
|
mResource->AppendData(aData);
|
|
nsTArray<MediaByteRange> byteRanges;
|
|
MediaByteRange mbr =
|
|
MediaByteRange(mParser->mOffset, mResource->GetLength());
|
|
byteRanges.AppendElement(mbr);
|
|
mParser->RebuildFragmentedIndex(byteRanges);
|
|
|
|
if (initSegment || !HasCompleteInitData()) {
|
|
const MediaByteRange& range = mParser->mInitRange;
|
|
uint32_t length = range.mEnd - range.mStart;
|
|
if (length) {
|
|
if (!mInitData->SetLength(length)) {
|
|
// Super unlikely OOM
|
|
return false;
|
|
}
|
|
char* buffer = reinterpret_cast<char*>(mInitData->Elements());
|
|
mResource->ReadFromCache(buffer, range.mStart, length);
|
|
MSE_DEBUG(MP4ContainerParser ,"Stashed init of %u bytes.",
|
|
length);
|
|
} else {
|
|
MSE_DEBUG(MP4ContainerParser, "Incomplete init found.");
|
|
}
|
|
mHasInitData = true;
|
|
}
|
|
|
|
mp4_demuxer::Interval<mp4_demuxer::Microseconds> compositionRange =
|
|
mParser->GetCompositionRange(byteRanges);
|
|
mResource->EvictData(mParser->mOffset, mParser->mOffset);
|
|
|
|
if (compositionRange.IsNull()) {
|
|
return false;
|
|
}
|
|
aStart = compositionRange.start;
|
|
aEnd = compositionRange.end;
|
|
MSE_DEBUG(MP4ContainerParser, "[%lld, %lld]",
|
|
aStart, aEnd);
|
|
return true;
|
|
}
|
|
|
|
// Gaps of up to 20ms (marginally longer than a single frame at 60fps) are considered
|
|
// to be sequential frames.
|
|
int64_t GetRoundingError()
|
|
{
|
|
return 20000;
|
|
}
|
|
|
|
private:
|
|
nsRefPtr<MP4Stream> mStream;
|
|
nsAutoPtr<mp4_demuxer::MoofParser> mParser;
|
|
Monitor mMonitor;
|
|
};
|
|
#endif
|
|
|
|
/*static*/ ContainerParser*
|
|
ContainerParser::CreateForMIMEType(const nsACString& aType)
|
|
{
|
|
if (aType.LowerCaseEqualsLiteral("video/webm") || aType.LowerCaseEqualsLiteral("audio/webm")) {
|
|
return new WebMContainerParser(aType);
|
|
}
|
|
|
|
#ifdef MOZ_FMP4
|
|
if (aType.LowerCaseEqualsLiteral("video/mp4") || aType.LowerCaseEqualsLiteral("audio/mp4")) {
|
|
return new MP4ContainerParser(aType);
|
|
}
|
|
#endif
|
|
return new ContainerParser(aType);
|
|
}
|
|
|
|
#undef MSE_DEBUG
|
|
|
|
} // namespace mozilla
|