Files
UnrealEngineUWP/Engine/Source/Runtime/Android/AndroidAudio/Private/AndroidAudioBuffer.cpp
Matthew Griffin 98a1cdce0f Added Structs to store streamed audio chunks for runtime streaming or in DDC.
Followed pattern set out by texture streaming so that each chunk of audio data resides in its own bulkdata struct. It is currently possible to split audio into chunks when SoundWaves are marked for streaming but there is no way of doing this exposed at present.
Changed the parameters of FAudioDevice::GetRuntimeFormat so that the relevant SoundWave must be passed in, to allow for different formats for individual sounds/streaming options.
USoundWave::FreeResources no longer resets the NumChannels as it is unnecessary and causes sounds to be unable to play after the OGG data is flushed when attempting to switch to OPUS.

[CL 2099012 by Matthew Griffin in Main branch]
2014-06-09 11:13:16 -04:00

233 lines
6.6 KiB
C++

// Copyright 1998-2014 Epic Games, Inc. All Rights Reserved.
#include "AndroidAudioDevice.h"
#include "AudioEffect.h"
#include "Engine.h"
#include "IAudioFormat.h"
#include "AudioDecompress.h"
/*------------------------------------------------------------------------------------
FSLESSoundBuffer.
------------------------------------------------------------------------------------*/
/**
* Constructor
*
* @param AudioDevice audio device this sound buffer is going to be attached to.
*/
FSLESSoundBuffer::FSLESSoundBuffer( FSLESAudioDevice* InAudioDevice ) :
AudioDevice(InAudioDevice),
AudioData(NULL),
DecompressionState( NULL ),
Format(SoundFormat_Invalid)
{
}
/**
* Destructor
*
* Frees wave data and detaches itself from audio device.
*/
FSLESSoundBuffer::~FSLESSoundBuffer( void )
{
if( ResourceID )
{
AudioDevice->WaveBufferMap.Remove( ResourceID );
}
FMemory::Free( AudioData);
if( DecompressionState )
{
delete DecompressionState;
}
}
FSLESSoundBuffer* FSLESSoundBuffer::CreateQueuedBuffer( FSLESAudioDevice* AudioDevice, USoundWave* InWave )
{
// Always create a new buffer for real time decompressed sounds
FSLESSoundBuffer* Buffer = new FSLESSoundBuffer( AudioDevice);
// Prime the first two buffers and prepare the decompression
FSoundQualityInfo QualityInfo = { 0 };
Buffer->DecompressionState = AudioDevice->CreateCompressedAudioInfo(InWave);
InWave->InitAudioResource( AudioDevice->GetRuntimeFormat(InWave) );
if( Buffer->DecompressionState->ReadCompressedInfo( InWave->ResourceData, InWave->ResourceSize, &QualityInfo ) )
{
// Refresh the wave data
InWave->SampleRate = QualityInfo.SampleRate;
InWave->NumChannels = QualityInfo.NumChannels;
InWave->RawPCMDataSize = QualityInfo.SampleDataSize;
InWave->Duration = QualityInfo.Duration;
// Clear out any dangling pointers
Buffer->AudioData = NULL;
Buffer->BufferSize = 0;
// Keep track of associated resource name.
Buffer->ResourceName = InWave->GetPathName();
Buffer->NumChannels = InWave->NumChannels;
Buffer->SampleRate = InWave->SampleRate;
//Android can't handle more than 48kHz, so turn on halfrate decoding and adjust parameters
if (Buffer->SampleRate > 48000)
{
UE_LOG(LogAndroidAudio, Log, TEXT( "Converting %s to halfrate from %d" ), *InWave->GetName(), Buffer->SampleRate );
Buffer->DecompressionState->EnableHalfRate( true);
Buffer->SampleRate = Buffer->SampleRate / 2;
InWave->SampleRate = InWave->SampleRate / 2;
uint32 SampleCount = QualityInfo.SampleDataSize / (QualityInfo.NumChannels * sizeof(uint16));
SampleCount /= 2;
InWave->RawPCMDataSize = SampleCount * QualityInfo.NumChannels * sizeof(uint16);;
}
Buffer->Format = SoundFormat_PCMRT;
}
else
{
InWave->DecompressionType = DTYPE_Invalid;
InWave->NumChannels = 0;
InWave->RemoveAudioResource();
}
return( Buffer );
}
/**
* Static function used to create an OpenSL buffer and upload decompressed ogg vorbis data to.
*
* @param InWave USoundWave to use as template and wave source
* @param AudioDevice audio device to attach created buffer to
* @return FSLESSoundBuffer pointer if buffer creation succeeded, NULL otherwise
*/
FSLESSoundBuffer* FSLESSoundBuffer::CreateNativeBuffer( FSLESAudioDevice* AudioDevice, USoundWave* InWave )
{
#if WITH_OGGVORBIS
// Check to see if thread has finished decompressing on the other thread
if( InWave->AudioDecompressor != NULL )
{
InWave->AudioDecompressor->EnsureCompletion();
// Remove the decompressor
delete InWave->AudioDecompressor;
InWave->AudioDecompressor = NULL;
}
#endif //WITH_OGGVORBIS
FSLESSoundBuffer* Buffer = NULL;
// Create new buffer.
Buffer = new FSLESSoundBuffer( AudioDevice );
// Allocate new resource ID and assign to USoundNodeWave. A value of 0 (default) means not yet registered.
int32 ResourceID = AudioDevice->NextResourceID++;
Buffer->ResourceID = ResourceID;
InWave->ResourceID = ResourceID;
AudioDevice->Buffers.Add( Buffer );
AudioDevice->WaveBufferMap.Add( ResourceID, Buffer );
// Keep track of associated resource name.
Buffer->ResourceName = InWave->GetPathName();
Buffer->NumChannels = InWave->NumChannels;
Buffer->SampleRate = InWave->SampleRate;
// Take ownership the PCM data
Buffer->AudioData = InWave->RawPCMData;
Buffer->BufferSize = InWave->RawPCMDataSize;
Buffer->Format = SoundFormat_PCM;
InWave->RawPCMData = NULL;
InWave->RemoveAudioResource();
return Buffer;
}
/**
* Static function used to create a buffer.
*
* @param InWave USoundNodeWave to use as template and wave source
* @param AudioDevice audio device to attach created buffer to
* @param bIsPrecacheRequest Whether this request is for precaching or not
* @return FSLESSoundBuffer pointer if buffer creation succeeded, NULL otherwise
*/
FSLESSoundBuffer* FSLESSoundBuffer::Init( FSLESAudioDevice* AudioDevice ,USoundWave* InWave )
{
SCOPE_CYCLE_COUNTER( STAT_AudioResourceCreationTime );
// Can't create a buffer without any source data
if( InWave == NULL || InWave->NumChannels == 0 )
{
UE_LOG( LogAndroidAudio, Warning, TEXT("InitBuffer with Null sound data"));
return( NULL );
}
FSLESSoundBuffer* Buffer = NULL;
EDecompressionType DecompressionType = InWave->DecompressionType;
switch( DecompressionType )
{
case DTYPE_Setup:
// Has circumvented precache mechanism - precache now
AudioDevice->Precache(InWave, true, false);
// if it didn't change, we will recurse forever
check(InWave->DecompressionType != DTYPE_Setup);
// Recall this function with new decompression type
return( Init( AudioDevice, InWave ) );
break;
case DTYPE_Native:
// Upload entire wav
if( InWave->ResourceID )
{
Buffer = AudioDevice->WaveBufferMap.FindRef( InWave->ResourceID );
}
if( Buffer == NULL )
{
Buffer = CreateNativeBuffer( AudioDevice, InWave );
}
break;
case DTYPE_RealTime:
// Always create a new buffer for streaming ogg vorbis data
Buffer = CreateQueuedBuffer( AudioDevice, InWave );
break;
case DTYPE_Invalid:
case DTYPE_Preview:
case DTYPE_Procedural:
default:
UE_LOG( LogAndroidAudio, Warning, TEXT("Init Buffer on unsupported sound type name = %s type = %d"), *InWave->GetName(), int32(DecompressionType));
break;
}
return Buffer;
}
/**
* Decompresses a chunk of compressed audio to the destination memory
*
* @param Destination Memory to decompress to
* @param bLooping Whether to loop the sound seamlessly, or pad with zeroes
* @return Whether the sound looped or not
*/
bool FSLESSoundBuffer::ReadCompressedData( uint8* Destination, bool bLooping )
{
ensure( DecompressionState);
return( DecompressionState->ReadCompressedData( Destination, bLooping, MONO_PCM_BUFFER_SIZE * NumChannels ) );
}