Files
UnrealEngineUWP/Engine/Source/Runtime/Android/AndroidAudio/Private/AndroidAudioBuffer.cpp
Ben Marsh 4ba423868f Copying //UE4/Dev-Build to //UE4/Dev-Main (Source: //UE4/Dev-Build @ 3209340)
#lockdown Nick.Penwarden
#rb none

==========================
MAJOR FEATURES + CHANGES
==========================

Change 3209340 on 2016/11/23 by Ben.Marsh

	Convert UE4 codebase to an "include what you use" model - where every header just includes the dependencies it needs, rather than every source file including large monolithic headers like Engine.h and UnrealEd.h.

	Measured full rebuild times around 2x faster using XGE on Windows, and improvements of 25% or more for incremental builds and full rebuilds on most other platforms.

	  * Every header now includes everything it needs to compile.
	        * There's a CoreMinimal.h header that gets you a set of ubiquitous types from Core (eg. FString, FName, TArray, FVector, etc...). Most headers now include this first.
	        * There's a CoreTypes.h header that sets up primitive UE4 types and build macros (int32, PLATFORM_WIN64, etc...). All headers in Core include this first, as does CoreMinimal.h.
	  * Every .cpp file includes its matching .h file first.
	        * This helps validate that each header is including everything it needs to compile.
	  * No engine code includes a monolithic header such as Engine.h or UnrealEd.h any more.
	        * You will get a warning if you try to include one of these from the engine. They still exist for compatibility with game projects and do not produce warnings when included there.
	        * There have only been minor changes to our internal games down to accommodate these changes. The intent is for this to be as seamless as possible.
	  * No engine code explicitly includes a precompiled header any more.
	        * We still use PCHs, but they're force-included on the compiler command line by UnrealBuildTool instead. This lets us tune what they contain without breaking any existing include dependencies.
	        * PCHs are generated by a tool to get a statistical amount of coverage for the source files using it, and I've seeded the new shared PCHs to contain any header included by > 15% of source files.

	Tool used to generate this transform is at Engine\Source\Programs\IncludeTool.

[CL 3209342 by Ben Marsh in Main branch]
2016-11-23 15:48:37 -05:00

282 lines
8.2 KiB
C++

// Copyright 1998-2016 Epic Games, Inc. All Rights Reserved.
#include "AndroidAudioDevice.h"
#include "AudioEffect.h"
#include "IAudioFormat.h"
#include "AudioDecompress.h"
#include "ContentStreaming.h"
/*------------------------------------------------------------------------------------
FSLESSoundBuffer.
------------------------------------------------------------------------------------*/
/**
* Constructor
*
* @param AudioDevice audio device this sound buffer is going to be attached to.
*/
FSLESSoundBuffer::FSLESSoundBuffer( FSLESAudioDevice* InAudioDevice )
: FSoundBuffer(InAudioDevice),
AudioData(NULL),
DecompressionState( NULL ),
Format(SoundFormat_Invalid)
{
}
/**
* Destructor
*
* Frees wave data and detaches itself from audio device.
*/
FSLESSoundBuffer::~FSLESSoundBuffer( void )
{
FMemory::Free( AudioData);
if( DecompressionState )
{
delete DecompressionState;
}
}
FSLESSoundBuffer* FSLESSoundBuffer::CreateQueuedBuffer( FSLESAudioDevice* AudioDevice, USoundWave* InWave )
{
// Check to see if thread has finished decompressing on the other thread
if (InWave->AudioDecompressor != nullptr)
{
InWave->AudioDecompressor->EnsureCompletion();
// Remove the decompressor
delete InWave->AudioDecompressor;
InWave->AudioDecompressor = nullptr;
}
// Always create a new buffer for real time decompressed sounds
FSLESSoundBuffer* Buffer = new FSLESSoundBuffer( AudioDevice);
// Prime the first two buffers and prepare the decompression
FSoundQualityInfo QualityInfo = { 0 };
Buffer->DecompressionState = AudioDevice->CreateCompressedAudioInfo(InWave);
// If the buffer was precached as native, the resource data will have been lost and we need to re-initialize it
if (InWave->ResourceData == nullptr)
{
InWave->InitAudioResource(AudioDevice->GetRuntimeFormat(InWave));
}
if (Buffer->DecompressionState && Buffer->DecompressionState->ReadCompressedInfo(InWave->ResourceData, InWave->ResourceSize, &QualityInfo))
{
// Clear out any dangling pointers
Buffer->AudioData = NULL;
Buffer->BufferSize = 0;
// Keep track of associated resource name.
Buffer->ResourceName = InWave->GetPathName();
Buffer->NumChannels = InWave->NumChannels;
Buffer->SampleRate = InWave->SampleRate;
//Android can't handle more than 48kHz, so turn on halfrate decoding and adjust parameters
if (Buffer->SampleRate > 48000)
{
UE_LOG(LogAndroidAudio, Log, TEXT( "Converting %s to halfrate from %d" ), *InWave->GetName(), Buffer->SampleRate );
Buffer->DecompressionState->EnableHalfRate( true);
Buffer->SampleRate = Buffer->SampleRate / 2;
InWave->SampleRate = InWave->SampleRate / 2;
uint32 SampleCount = QualityInfo.SampleDataSize / (QualityInfo.NumChannels * sizeof(uint16));
SampleCount /= 2;
InWave->RawPCMDataSize = SampleCount * QualityInfo.NumChannels * sizeof(uint16);;
}
Buffer->Format = SoundFormat_PCMRT;
}
else
{
InWave->DecompressionType = DTYPE_Invalid;
InWave->NumChannels = 0;
InWave->RemoveAudioResource();
}
return Buffer;
}
/**
* Static function used to create an OpenSL buffer and upload decompressed ogg vorbis data to.
*
* @param InWave USoundWave to use as template and wave source
* @param AudioDevice audio device to attach created buffer to
* @return FSLESSoundBuffer pointer if buffer creation succeeded, NULL otherwise
*/
FSLESSoundBuffer* FSLESSoundBuffer::CreateNativeBuffer( FSLESAudioDevice* AudioDevice, USoundWave* InWave )
{
// Check to see if thread has finished decompressing on the other thread
if( InWave->AudioDecompressor != NULL )
{
InWave->AudioDecompressor->EnsureCompletion();
// Remove the decompressor
delete InWave->AudioDecompressor;
InWave->AudioDecompressor = NULL;
}
FSLESSoundBuffer* Buffer = NULL;
// Create new buffer.
Buffer = new FSLESSoundBuffer( AudioDevice );
Buffer->DecompressionState = AudioDevice->CreateCompressedAudioInfo(InWave);
FAudioDeviceManager* AudioDeviceManager = GEngine->GetAudioDeviceManager();
check(AudioDeviceManager != nullptr);
AudioDeviceManager->TrackResource(InWave, Buffer);
Buffer->NumChannels = InWave->NumChannels;
Buffer->SampleRate = InWave->SampleRate;
// Take ownership the PCM data
Buffer->AudioData = InWave->RawPCMData;
Buffer->BufferSize = InWave->RawPCMDataSize;
Buffer->Format = SoundFormat_PCM;
InWave->RawPCMData = NULL;
InWave->RemoveAudioResource();
return Buffer;
}
/**
* Static function used to create an Audio buffer and dynamically upload procedural data to.
*
* @param InWave USoundWave to use as template and wave source
* @param AudioDevice audio device to attach created buffer to
* @return FSLESSoundBuffer pointer if buffer creation succeeded, NULL otherwise
*/
FSLESSoundBuffer* FSLESSoundBuffer::CreateProceduralBuffer(FSLESAudioDevice* AudioDevice, USoundWave* InWave)
{
FSLESSoundBuffer* Buffer = new FSLESSoundBuffer(AudioDevice);
// Setup any default information
Buffer->DecompressionState = NULL;
Buffer->AudioData = NULL;
Buffer->BufferSize = 0;
Buffer->Format = SoundFormat_PCMRT;
Buffer->NumChannels = InWave->NumChannels;
Buffer->SampleRate = InWave->SampleRate;
InWave->RawPCMData = NULL;
// No tracking of this resource as it's temporary
Buffer->ResourceID = 0;
InWave->ResourceID = 0;
return Buffer;
}
/**
* Static function used to create a buffer.
*
* @param InWave USoundNodeWave to use as template and wave source
* @param AudioDevice audio device to attach created buffer to
* @param bIsPrecacheRequest Whether this request is for precaching or not
* @return FSLESSoundBuffer pointer if buffer creation succeeded, NULL otherwise
*/
FSLESSoundBuffer* FSLESSoundBuffer::Init( FSLESAudioDevice* AudioDevice ,USoundWave* InWave )
{
SCOPE_CYCLE_COUNTER( STAT_AudioResourceCreationTime );
// Can't create a buffer without any source data
if( InWave == NULL || InWave->NumChannels == 0 )
{
UE_LOG( LogAndroidAudio, Warning, TEXT("InitBuffer with Null sound data"));
return( NULL );
}
FAudioDeviceManager* AudioDeviceManager = GEngine->GetAudioDeviceManager();
FSLESSoundBuffer* Buffer = NULL;
EDecompressionType DecompressionType = InWave->DecompressionType;
UE_LOG(LogAndroidAudio, Verbose, TEXT("Init: Using decompression type: %d"), int32(DecompressionType));
switch( DecompressionType )
{
case DTYPE_Setup:
// Has circumvented precache mechanism - precache now
AudioDevice->Precache(InWave, true, false);
// if it didn't change, we will recurse forever
check(InWave->DecompressionType != DTYPE_Setup);
// Recall this function with new decompression type
return( Init( AudioDevice, InWave ) );
break;
case DTYPE_Native:
// Upload entire wav
if( InWave->ResourceID )
{
Buffer = static_cast<FSLESSoundBuffer*>(AudioDeviceManager->WaveBufferMap.FindRef(InWave->ResourceID));
}
if( Buffer == NULL )
{
Buffer = CreateNativeBuffer( AudioDevice, InWave );
}
break;
case DTYPE_RealTime:
// Always create a new buffer for streaming ogg vorbis data
Buffer = CreateQueuedBuffer( AudioDevice, InWave );
break;
case DTYPE_Procedural:
// New buffer for procedural data
Buffer = CreateProceduralBuffer(AudioDevice, InWave);
break;
case DTYPE_Invalid:
case DTYPE_Preview:
default:
UE_LOG( LogAndroidAudio, Warning, TEXT("Init Buffer on unsupported sound type name = %s type = %d"), *InWave->GetName(), int32(DecompressionType));
break;
}
return Buffer;
}
/**
* Decompresses a chunk of compressed audio to the destination memory
*
* @param Destination Memory to decompress to
* @param bLooping Whether to loop the sound seamlessly, or pad with zeroes
* @return Whether the sound looped or not
*/
bool FSLESSoundBuffer::ReadCompressedData( uint8* Destination, bool bLooping )
{
ensure( DecompressionState);
return(DecompressionState->ReadCompressedData(Destination, bLooping, DecompressionState->GetStreamBufferSize() * NumChannels));
}
void FSLESSoundBuffer::Seek(const float SeekTime)
{
if (ensure(DecompressionState))
{
DecompressionState->SeekToTime(SeekTime);
}
}
/**
* Returns the size for a real time/streaming buffer based on decompressor
*
* @return Size of buffer in bytes for a single channel or 0 if no decompression state
*/
int FSLESSoundBuffer::GetRTBufferSize(void)
{
return DecompressionState ? DecompressionState->GetStreamBufferSize() : MONO_PCM_BUFFER_SIZE;
}