You've already forked UnrealEngineUWP
mirror of
https://github.com/izzy2lost/UnrealEngineUWP.git
synced 2026-03-26 18:15:20 -07:00
#jira UE-187406, UE-187390, UE-187404, UE-187403, UE-187405, UE-187392, UE-187391, UE-187395, UE-187399, UE-187398, UE-187389, UE-187393, UE-187394, UE-187396, UE-187397, UE-187401 #rb phil.popp [CL 26130575 by Maxwell Hayes in 5.3 branch]
314 lines
11 KiB
C++
314 lines
11 KiB
C++
// Copyright Epic Games, Inc. All Rights Reserved.
|
|
|
|
#include "AudioMixerDevice.h"
|
|
#include "AudioBusSubsystem.h"
|
|
#include "AudioDevice.h"
|
|
#include "Internationalization/Text.h"
|
|
#include "MediaPacket.h"
|
|
#include "MetasoundAudioBuffer.h"
|
|
#include "MetasoundAudioBus.h"
|
|
#include "MetasoundEngineNodesNames.h"
|
|
#include "MetasoundExecutableOperator.h"
|
|
#include "MetasoundFacade.h"
|
|
#include "MetasoundNodeRegistrationMacro.h"
|
|
#include "MetasoundParamHelper.h"
|
|
#include "MetasoundStandardNodesCategories.h"
|
|
|
|
#define LOCTEXT_NAMESPACE "MetasoundAudioBusWriterNode"
|
|
|
|
namespace Metasound
|
|
{
|
|
namespace AudioBusWriterNode
|
|
{
|
|
METASOUND_PARAM(InParamAudioBusOutput, "Audio Bus", "Audio Bus Asset.");
|
|
METASOUND_PARAM(InParamAudio, "In {0}", "Audio input for channel {0}.");
|
|
}
|
|
|
|
template<uint32 NumChannels>
|
|
class TAudioBusWriterOperator : public TExecutableOperator<TAudioBusWriterOperator<NumChannels>>
|
|
{
|
|
public:
|
|
static const FNodeClassMetadata& GetNodeInfo()
|
|
{
|
|
auto InitNodeInfo = []() -> FNodeClassMetadata
|
|
{
|
|
FName OperatorName = *FString::Printf(TEXT("Audio Bus Writer (%d)"), NumChannels);
|
|
FText NodeDisplayName = METASOUND_LOCTEXT_FORMAT("AudioBusWriterDisplayNamePattern", "Audio Bus Writer ({0})", NumChannels);
|
|
|
|
FNodeClassMetadata Info;
|
|
Info.ClassName = { EngineNodes::Namespace, OperatorName, TEXT("") };
|
|
Info.MajorVersion = 1;
|
|
Info.MinorVersion = 0;
|
|
Info.DisplayName = NodeDisplayName;
|
|
Info.Description = METASOUND_LOCTEXT("AudioBusWriter_Description", "Sends audio data to the audio bus asset.");
|
|
Info.Author = PluginAuthor;
|
|
Info.PromptIfMissing = PluginNodeMissingPrompt;
|
|
Info.DefaultInterface = GetVertexInterface();
|
|
Info.CategoryHierarchy.Emplace(NodeCategories::Io);
|
|
return Info;
|
|
};
|
|
|
|
static const FNodeClassMetadata Info = InitNodeInfo();
|
|
|
|
return Info;
|
|
}
|
|
|
|
static const FVertexInterface& GetVertexInterface()
|
|
{
|
|
using namespace AudioBusWriterNode;
|
|
|
|
auto CreateVertexInterface = []() -> FVertexInterface
|
|
{
|
|
FInputVertexInterface InputInterface;
|
|
InputInterface.Add(TInputDataVertex<FAudioBusAsset>(METASOUND_GET_PARAM_NAME_AND_METADATA(InParamAudioBusOutput)));
|
|
for (uint32 i = 0; i < NumChannels; ++i)
|
|
{
|
|
InputInterface.Add(TInputDataVertex<FAudioBuffer>(METASOUND_GET_PARAM_NAME_WITH_INDEX_AND_METADATA(InParamAudio, i)));
|
|
}
|
|
|
|
FOutputVertexInterface OutputInterface;
|
|
|
|
return FVertexInterface(InputInterface, OutputInterface);
|
|
};
|
|
|
|
static const FVertexInterface Interface = CreateVertexInterface();
|
|
return Interface;
|
|
}
|
|
|
|
static TUniquePtr<IOperator> CreateOperator(const FCreateOperatorParams& InParams, FBuildErrorArray& OutErrors)
|
|
{
|
|
using namespace Frontend;
|
|
using namespace AudioBusWriterNode;
|
|
|
|
const FDataReferenceCollection& InputCollection = InParams.InputDataReferences;
|
|
|
|
bool bHasEnvironmentVars = InParams.Environment.Contains<Audio::FDeviceId>(SourceInterface::Environment::DeviceID);
|
|
bHasEnvironmentVars &= InParams.Environment.Contains<int32>(SourceInterface::Environment::AudioMixerNumOutputFrames);
|
|
|
|
if (bHasEnvironmentVars)
|
|
{
|
|
FAudioBusAssetReadRef AudioBusIn = InputCollection.GetDataReadReferenceOrConstruct<FAudioBusAsset>(METASOUND_GET_PARAM_NAME(InParamAudioBusOutput));
|
|
|
|
TArray<FAudioBufferReadRef> AudioInputs;
|
|
for (int32 ChannelIndex = 0; ChannelIndex < NumChannels; ++ChannelIndex)
|
|
{
|
|
AudioInputs.Add(InputCollection.GetDataReadReferenceOrConstruct<FAudioBuffer>(METASOUND_GET_PARAM_NAME_WITH_INDEX(InParamAudio, ChannelIndex), InParams.OperatorSettings));
|
|
}
|
|
|
|
return MakeUnique<TAudioBusWriterOperator<NumChannels>>(InParams, MoveTemp(AudioBusIn), MoveTemp(AudioInputs));
|
|
}
|
|
else
|
|
{
|
|
UE_LOG(LogMetaSound, Warning, TEXT("Audio bus writer node requires audio device ID '%s' and audio mixer num output frames '%s' environment variables")
|
|
, *SourceInterface::Environment::DeviceID.ToString(), *SourceInterface::Environment::AudioMixerNumOutputFrames.ToString());
|
|
return nullptr;
|
|
}
|
|
}
|
|
|
|
TAudioBusWriterOperator(const FCreateOperatorParams& InParams, FAudioBusAssetReadRef InAudioBusAsset, TArray<FAudioBufferReadRef> InAudioInputs)
|
|
: AudioBusAsset(MoveTemp(InAudioBusAsset))
|
|
, AudioInputs(MoveTemp(InAudioInputs))
|
|
{
|
|
Reset(InParams);
|
|
}
|
|
|
|
void CreatePatchInput()
|
|
{
|
|
const FAudioBusProxyPtr& AudioBusProxy = AudioBusAsset->GetAudioBusProxy();
|
|
if (AudioBusProxy.IsValid())
|
|
{
|
|
if (FAudioDeviceManager* ADM = FAudioDeviceManager::Get())
|
|
{
|
|
if (FAudioDevice* AudioDevice = ADM->GetAudioDeviceRaw(AudioDeviceId))
|
|
{
|
|
UAudioBusSubsystem* AudioBusSubsystem = AudioDevice->GetSubsystem<UAudioBusSubsystem>();
|
|
check(AudioBusSubsystem);
|
|
|
|
AudioBusId = AudioBusProxy->AudioBusId;
|
|
const Audio::FAudioBusKey AudioBusKey = Audio::FAudioBusKey(AudioBusId);
|
|
|
|
// Start the audio bus in case it's not already started
|
|
AudioBusChannels = AudioBusProxy->NumChannels;
|
|
AudioBusSubsystem->StartAudioBus(AudioBusKey, AudioBusChannels, false);
|
|
|
|
InterleavedBuffer.Reset();
|
|
InterleavedBuffer.AddZeroed(BlockSizeFrames * AudioBusChannels);
|
|
|
|
// Create a bus patch input with enough room for the number of samples we expect and some buffering
|
|
AudioBusPatchInput = AudioBusSubsystem->AddPatchInputForAudioBus(AudioBusKey, BlockSizeFrames, AudioBusChannels);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void Reset(const IOperator::FResetParams& InParams)
|
|
{
|
|
using namespace Frontend;
|
|
using namespace AudioBusWriterNode;
|
|
|
|
bool bHasEnvironmentVars = InParams.Environment.Contains<Audio::FDeviceId>(SourceInterface::Environment::DeviceID);
|
|
bHasEnvironmentVars &= InParams.Environment.Contains<int32>(SourceInterface::Environment::AudioMixerNumOutputFrames);
|
|
|
|
if (bHasEnvironmentVars)
|
|
{
|
|
SampleRate = InParams.OperatorSettings.GetSampleRate();
|
|
AudioDeviceId = InParams.Environment.GetValue<Audio::FDeviceId>(SourceInterface::Environment::DeviceID);
|
|
AudioMixerOutputFrames = InParams.Environment.GetValue<int32>(SourceInterface::Environment::AudioMixerNumOutputFrames);
|
|
}
|
|
else
|
|
{
|
|
UE_LOG(LogMetaSound, Warning, TEXT("Audio bus writer node requires audio device ID '%s' and audio mixer num output frames '%s' environment variables")
|
|
, *SourceInterface::Environment::DeviceID.ToString(), *SourceInterface::Environment::AudioMixerNumOutputFrames.ToString());
|
|
}
|
|
|
|
BlockSizeFrames = InParams.OperatorSettings.GetNumFramesPerBlock();
|
|
|
|
CreatePatchInput();
|
|
bFirstBlock = true;
|
|
}
|
|
|
|
|
|
virtual void BindInputs(FInputVertexInterfaceData& InOutVertexData) override
|
|
{
|
|
using namespace AudioBusWriterNode;
|
|
|
|
InOutVertexData.BindReadVertex(METASOUND_GET_PARAM_NAME(InParamAudioBusOutput), AudioBusAsset);
|
|
|
|
for (int32 ChannelIndex = 0; ChannelIndex < NumChannels; ++ChannelIndex)
|
|
{
|
|
InOutVertexData.BindReadVertex(METASOUND_GET_PARAM_NAME_WITH_INDEX(InParamAudio, ChannelIndex), AudioInputs[ChannelIndex]);
|
|
}
|
|
}
|
|
|
|
virtual void BindOutputs(FOutputVertexInterfaceData& InOutVertexData) override
|
|
{
|
|
}
|
|
|
|
virtual FDataReferenceCollection GetInputs() const override
|
|
{
|
|
// This should never be called. Bind(...) is called instead. This method
|
|
// exists as a stop-gap until the API can be deprecated and removed.
|
|
checkNoEntry();
|
|
return {};
|
|
}
|
|
|
|
virtual FDataReferenceCollection GetOutputs() const override
|
|
{
|
|
// This should never be called. Bind(...) is called instead. This method
|
|
// exists as a stop-gap until the API can be deprecated and removed.
|
|
checkNoEntry();
|
|
return {};
|
|
}
|
|
|
|
void Execute()
|
|
{
|
|
const FAudioBusProxyPtr& BusProxy = AudioBusAsset->GetAudioBusProxy();
|
|
if (BusProxy.IsValid() && BusProxy->AudioBusId != AudioBusId)
|
|
{
|
|
AudioBusPatchInput.Reset();
|
|
}
|
|
|
|
if (!AudioBusPatchInput.IsValid())
|
|
{
|
|
// if environment vars & a valid audio bus have been set since starting, try to create the patch now
|
|
if (SampleRate > 0.f && BusProxy.IsValid())
|
|
{
|
|
CreatePatchInput();
|
|
}
|
|
else
|
|
{
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (!AudioBusPatchInput.IsOutputStillActive())
|
|
{
|
|
return;
|
|
}
|
|
|
|
// Retrieve input and interleaved buffer pointers
|
|
const float* AudioInputBufferPtrs[NumChannels];
|
|
for (uint32 ChannelIndex = 0; ChannelIndex < NumChannels; ++ChannelIndex)
|
|
{
|
|
AudioInputBufferPtrs[ChannelIndex] = AudioInputs[ChannelIndex]->GetData();
|
|
}
|
|
float* InterleavedBufferPtr = InterleavedBuffer.GetData();
|
|
|
|
if (AudioBusChannels == 1)
|
|
{
|
|
FMemory::Memcpy(InterleavedBufferPtr, AudioInputBufferPtrs[0], BlockSizeFrames * sizeof(float));
|
|
}
|
|
else
|
|
{
|
|
// Interleave the inputs
|
|
// Writing the channels of the interleaved buffer sequentially should improve
|
|
// cache utilization compared to writing each input's frames sequentially.
|
|
// There is more likely to be a cache line for each buffer than for the
|
|
// entirety of the interleaved buffer.
|
|
uint32 MinChannels = FMath::Min(AudioBusChannels, NumChannels);
|
|
for (int32 FrameIndex = 0; FrameIndex < BlockSizeFrames; ++FrameIndex)
|
|
{
|
|
for (uint32 ChannelIndex = 0; ChannelIndex < MinChannels; ++ChannelIndex)
|
|
{
|
|
InterleavedBufferPtr[ChannelIndex] = *AudioInputBufferPtrs[ChannelIndex]++;
|
|
}
|
|
InterleavedBufferPtr += AudioBusChannels;
|
|
}
|
|
}
|
|
|
|
if (bFirstBlock)
|
|
{
|
|
bFirstBlock = false;
|
|
if (AudioMixerOutputFrames != BlockSizeFrames)
|
|
{
|
|
// Ensure there will be enough samples in the patch input to support the maximum metasound executions the mixer requires to fill its output frames after the next push.
|
|
AudioBusPatchInput.PushAudio(nullptr, (FMath::DivideAndRoundUp(FMath::Max(AudioMixerOutputFrames, BlockSizeFrames), FMath::Min(AudioMixerOutputFrames, BlockSizeFrames)) - 1) * BlockSizeFrames * AudioBusChannels);
|
|
}
|
|
}
|
|
|
|
// Pushes the interleaved data to the audio bus
|
|
int32 SamplesPushed = AudioBusPatchInput.PushAudio(InterleavedBuffer.GetData(), InterleavedBuffer.Num());
|
|
if (SamplesPushed < InterleavedBuffer.Num())
|
|
{
|
|
UE_LOG(LogMetaSound, Warning, TEXT("Underrun detected in audio bus writer node."));
|
|
}
|
|
}
|
|
|
|
private:
|
|
FAudioBusAssetReadRef AudioBusAsset;
|
|
TArray<FAudioBufferReadRef> AudioInputs;
|
|
|
|
TArray<float> InterleavedBuffer;
|
|
int32 AudioMixerOutputFrames = INDEX_NONE;
|
|
Audio::FDeviceId AudioDeviceId = INDEX_NONE;
|
|
float SampleRate = 0.0f;
|
|
Audio::FPatchInput AudioBusPatchInput;
|
|
uint32 AudioBusChannels = INDEX_NONE;
|
|
uint32 AudioBusId = 0;
|
|
int32 BlockSizeFrames = 0;
|
|
bool bFirstBlock = true;
|
|
};
|
|
|
|
template<uint32 NumChannels>
|
|
class TAudioBusWriterNode : public FNodeFacade
|
|
{
|
|
public:
|
|
TAudioBusWriterNode(const FNodeInitData& InitData)
|
|
: FNodeFacade(InitData.InstanceName, InitData.InstanceID, TFacadeOperatorClass<TAudioBusWriterOperator<NumChannels>>())
|
|
{
|
|
}
|
|
};
|
|
|
|
#define REGISTER_AUDIO_BUS_WRITER_NODE(ChannelCount) \
|
|
using FAudioBusWriterNode_##ChannelCount = TAudioBusWriterNode<ChannelCount>; \
|
|
METASOUND_REGISTER_NODE(FAudioBusWriterNode_##ChannelCount) \
|
|
|
|
REGISTER_AUDIO_BUS_WRITER_NODE(1);
|
|
REGISTER_AUDIO_BUS_WRITER_NODE(2);
|
|
REGISTER_AUDIO_BUS_WRITER_NODE(4);
|
|
REGISTER_AUDIO_BUS_WRITER_NODE(6);
|
|
REGISTER_AUDIO_BUS_WRITER_NODE(8);
|
|
}
|
|
|
|
#undef LOCTEXT_NAMESPACE
|