You've already forked openal-soft
mirror of
https://github.com/OldUnreal/openal-soft.git
synced 2026-04-02 21:38:01 -07:00
458 lines
17 KiB
C++
458 lines
17 KiB
C++
|
|
#include "config.h"
|
|
|
|
#include "converter.h"
|
|
|
|
#include <algorithm>
|
|
#include <cassert>
|
|
#include <cmath>
|
|
#include <cstddef>
|
|
#include <cstdint>
|
|
#include <iterator>
|
|
#include <climits>
|
|
|
|
#include "albit.h"
|
|
#include "alnumeric.h"
|
|
#include "fpu_ctrl.h"
|
|
|
|
|
|
namespace {
|
|
|
|
constexpr uint MaxPitch{10};
|
|
|
|
static_assert((BufferLineSize-1)/MaxPitch > 0, "MaxPitch is too large for BufferLineSize!");
|
|
static_assert((INT_MAX>>MixerFracBits)/MaxPitch > BufferLineSize,
|
|
"MaxPitch and/or BufferLineSize are too large for MixerFracBits!");
|
|
|
|
template<DevFmtType T>
|
|
constexpr float LoadSample(DevFmtType_t<T> val) noexcept = delete;
|
|
|
|
template<> constexpr float LoadSample<DevFmtByte>(DevFmtType_t<DevFmtByte> val) noexcept
|
|
{ return float(val) * (1.0f/128.0f); }
|
|
template<> constexpr float LoadSample<DevFmtShort>(DevFmtType_t<DevFmtShort> val) noexcept
|
|
{ return float(val) * (1.0f/32768.0f); }
|
|
template<> constexpr float LoadSample<DevFmtInt>(DevFmtType_t<DevFmtInt> val) noexcept
|
|
{ return static_cast<float>(val) * (1.0f/2147483648.0f); }
|
|
template<> constexpr float LoadSample<DevFmtFloat>(DevFmtType_t<DevFmtFloat> val) noexcept
|
|
{ return val; }
|
|
|
|
template<> constexpr float LoadSample<DevFmtUByte>(DevFmtType_t<DevFmtUByte> val) noexcept
|
|
{ return LoadSample<DevFmtByte>(static_cast<int8_t>(val - 128)); }
|
|
template<> constexpr float LoadSample<DevFmtUShort>(DevFmtType_t<DevFmtUShort> val) noexcept
|
|
{ return LoadSample<DevFmtShort>(static_cast<int16_t>(val - 32768)); }
|
|
template<> constexpr float LoadSample<DevFmtUInt>(DevFmtType_t<DevFmtUInt> val) noexcept
|
|
{ return LoadSample<DevFmtInt>(static_cast<int32_t>(val - 2147483648u)); }
|
|
|
|
|
|
template<DevFmtType T>
|
|
inline void LoadSampleArray(const al::span<float> dst, const void *src, const size_t channel,
|
|
const size_t srcstep) noexcept
|
|
{
|
|
assert(channel < srcstep);
|
|
const auto srcspan = al::span{static_cast<const DevFmtType_t<T>*>(src), dst.size()*srcstep};
|
|
auto ssrc = srcspan.cbegin();
|
|
std::generate(dst.begin(), dst.end(), [&ssrc,channel,srcstep]
|
|
{
|
|
const float ret{LoadSample<T>(ssrc[channel])};
|
|
ssrc += ptrdiff_t(srcstep);
|
|
return ret;
|
|
});
|
|
}
|
|
|
|
void LoadSamples(const al::span<float> dst, const void *src, const size_t channel,
|
|
const size_t srcstep, const DevFmtType srctype) noexcept
|
|
{
|
|
#define HANDLE_FMT(T) \
|
|
case T: LoadSampleArray<T>(dst, src, channel, srcstep); break
|
|
switch(srctype)
|
|
{
|
|
HANDLE_FMT(DevFmtByte);
|
|
HANDLE_FMT(DevFmtUByte);
|
|
HANDLE_FMT(DevFmtShort);
|
|
HANDLE_FMT(DevFmtUShort);
|
|
HANDLE_FMT(DevFmtInt);
|
|
HANDLE_FMT(DevFmtUInt);
|
|
HANDLE_FMT(DevFmtFloat);
|
|
}
|
|
#undef HANDLE_FMT
|
|
}
|
|
|
|
|
|
template<DevFmtType T>
|
|
inline DevFmtType_t<T> StoreSample(float) noexcept;
|
|
|
|
template<> inline float StoreSample<DevFmtFloat>(float val) noexcept
|
|
{ return val; }
|
|
template<> inline int32_t StoreSample<DevFmtInt>(float val) noexcept
|
|
{ return fastf2i(std::clamp(val*2147483648.0f, -2147483648.0f, 2147483520.0f)); }
|
|
template<> inline int16_t StoreSample<DevFmtShort>(float val) noexcept
|
|
{ return static_cast<int16_t>(fastf2i(std::clamp(val*32768.0f, -32768.0f, 32767.0f))); }
|
|
template<> inline int8_t StoreSample<DevFmtByte>(float val) noexcept
|
|
{ return static_cast<int8_t>(fastf2i(std::clamp(val*128.0f, -128.0f, 127.0f))); }
|
|
|
|
/* Define unsigned output variations. */
|
|
template<> inline uint32_t StoreSample<DevFmtUInt>(float val) noexcept
|
|
{ return static_cast<uint32_t>(StoreSample<DevFmtInt>(val)) + 2147483648u; }
|
|
template<> inline uint16_t StoreSample<DevFmtUShort>(float val) noexcept
|
|
{ return static_cast<uint16_t>(StoreSample<DevFmtShort>(val) + 32768); }
|
|
template<> inline uint8_t StoreSample<DevFmtUByte>(float val) noexcept
|
|
{ return static_cast<uint8_t>(StoreSample<DevFmtByte>(val) + 128); }
|
|
|
|
template<DevFmtType T>
|
|
inline void StoreSampleArray(void *dst, const al::span<const float> src, const size_t channel,
|
|
const size_t dststep) noexcept
|
|
{
|
|
assert(channel < dststep);
|
|
const auto dstspan = al::span{static_cast<DevFmtType_t<T>*>(dst), src.size()*dststep};
|
|
auto sdst = dstspan.begin();
|
|
std::for_each(src.cbegin(), src.cend(), [&sdst,channel,dststep](const float in)
|
|
{
|
|
sdst[channel] = StoreSample<T>(in);
|
|
sdst += ptrdiff_t(dststep);
|
|
});
|
|
}
|
|
|
|
|
|
void StoreSamples(void *dst, const al::span<const float> src, const size_t channel,
|
|
const size_t dststep, const DevFmtType dsttype) noexcept
|
|
{
|
|
#define HANDLE_FMT(T) \
|
|
case T: StoreSampleArray<T>(dst, src, channel, dststep); break
|
|
switch(dsttype)
|
|
{
|
|
HANDLE_FMT(DevFmtByte);
|
|
HANDLE_FMT(DevFmtUByte);
|
|
HANDLE_FMT(DevFmtShort);
|
|
HANDLE_FMT(DevFmtUShort);
|
|
HANDLE_FMT(DevFmtInt);
|
|
HANDLE_FMT(DevFmtUInt);
|
|
HANDLE_FMT(DevFmtFloat);
|
|
}
|
|
#undef HANDLE_FMT
|
|
}
|
|
|
|
|
|
template<DevFmtType T>
|
|
void Mono2Stereo(const al::span<float> dst, const void *src) noexcept
|
|
{
|
|
const auto srcspan = al::span{static_cast<const DevFmtType_t<T>*>(src), dst.size()>>1};
|
|
auto sdst = dst.begin();
|
|
std::for_each(srcspan.cbegin(), srcspan.cend(), [&sdst](const auto in)
|
|
{ sdst = std::fill_n(sdst, 2, LoadSample<T>(in)*0.707106781187f); });
|
|
}
|
|
|
|
template<DevFmtType T>
|
|
void Multi2Mono(uint chanmask, const size_t step, const float scale, const al::span<float> dst,
|
|
const void *src) noexcept
|
|
{
|
|
const auto srcspan = al::span{static_cast<const DevFmtType_t<T>*>(src), step*dst.size()};
|
|
std::fill_n(dst.begin(), dst.size(), 0.0f);
|
|
for(size_t c{0};chanmask;++c)
|
|
{
|
|
if((chanmask&1)) LIKELY
|
|
{
|
|
auto ssrc = srcspan.cbegin();
|
|
std::for_each(dst.begin(), dst.end(), [&ssrc,step,c](float &sample)
|
|
{
|
|
const float s{LoadSample<T>(ssrc[c])};
|
|
ssrc += ptrdiff_t(step);
|
|
sample += s;
|
|
});
|
|
}
|
|
chanmask >>= 1;
|
|
}
|
|
std::for_each(dst.begin(), dst.end(), [scale](float &sample) noexcept { sample *= scale; });
|
|
}
|
|
|
|
} // namespace
|
|
|
|
SampleConverterPtr SampleConverter::Create(DevFmtType srcType, DevFmtType dstType, size_t numchans,
|
|
uint srcRate, uint dstRate, Resampler resampler)
|
|
{
|
|
if(numchans < 1 || srcRate < 1 || dstRate < 1)
|
|
return nullptr;
|
|
|
|
SampleConverterPtr converter{new(FamCount(numchans)) SampleConverter{numchans}};
|
|
converter->mSrcType = srcType;
|
|
converter->mDstType = dstType;
|
|
converter->mSrcTypeSize = BytesFromDevFmt(srcType);
|
|
converter->mDstTypeSize = BytesFromDevFmt(dstType);
|
|
|
|
converter->mSrcPrepCount = MaxResamplerPadding;
|
|
converter->mFracOffset = 0;
|
|
for(auto &chan : converter->mChan)
|
|
chan.PrevSamples.fill(0.0f);
|
|
|
|
/* Have to set the mixer FPU mode since that's what the resampler code expects. */
|
|
FPUCtl mixer_mode{};
|
|
const auto step = std::min(std::round(srcRate*double{MixerFracOne}/dstRate),
|
|
MaxPitch*double{MixerFracOne});
|
|
converter->mIncrement = std::max(static_cast<uint>(step), 1u);
|
|
if(converter->mIncrement == MixerFracOne)
|
|
{
|
|
converter->mResample = [](const InterpState*, const al::span<const float> src, uint,
|
|
const uint, const al::span<float> dst)
|
|
{ std::copy_n(src.begin()+MaxResamplerEdge, dst.size(), dst.begin()); };
|
|
}
|
|
else
|
|
converter->mResample = PrepareResampler(resampler, converter->mIncrement,
|
|
&converter->mState);
|
|
|
|
return converter;
|
|
}
|
|
|
|
uint SampleConverter::availableOut(uint srcframes) const
|
|
{
|
|
if(srcframes < 1)
|
|
{
|
|
/* No output samples if there's no input samples. */
|
|
return 0;
|
|
}
|
|
|
|
const uint prepcount{mSrcPrepCount};
|
|
if(prepcount < MaxResamplerPadding && MaxResamplerPadding - prepcount >= srcframes)
|
|
{
|
|
/* Not enough input samples to generate an output sample. */
|
|
return 0;
|
|
}
|
|
|
|
uint64_t DataSize64{prepcount};
|
|
DataSize64 += srcframes;
|
|
DataSize64 -= MaxResamplerPadding;
|
|
DataSize64 <<= MixerFracBits;
|
|
DataSize64 -= mFracOffset;
|
|
|
|
/* If we have a full prep, we can generate at least one sample. */
|
|
return static_cast<uint>(std::clamp((DataSize64 + mIncrement-1)/mIncrement, 1_u64,
|
|
uint64_t{std::numeric_limits<int>::max()}));
|
|
}
|
|
|
|
uint SampleConverter::convert(const void **src, uint *srcframes, void *dst, uint dstframes)
|
|
{
|
|
const size_t SrcFrameSize{mChan.size() * mSrcTypeSize};
|
|
const size_t DstFrameSize{mChan.size() * mDstTypeSize};
|
|
const uint increment{mIncrement};
|
|
uint NumSrcSamples{*srcframes};
|
|
auto SamplesIn = al::span{static_cast<const std::byte*>(*src), NumSrcSamples*SrcFrameSize};
|
|
auto SamplesOut = al::span{static_cast<std::byte*>(dst), dstframes*DstFrameSize};
|
|
|
|
FPUCtl mixer_mode{};
|
|
uint pos{0};
|
|
while(pos < dstframes && NumSrcSamples > 0)
|
|
{
|
|
const uint prepcount{mSrcPrepCount};
|
|
const uint readable{std::min(NumSrcSamples, uint{BufferLineSize} - prepcount)};
|
|
|
|
if(prepcount < MaxResamplerPadding && MaxResamplerPadding-prepcount >= readable)
|
|
{
|
|
/* Not enough input samples to generate an output sample. Store
|
|
* what we're given for later.
|
|
*/
|
|
for(size_t chan{0u};chan < mChan.size();chan++)
|
|
LoadSamples(al::span{mChan[chan].PrevSamples}.subspan(prepcount, readable),
|
|
SamplesIn.data(), chan, mChan.size(), mSrcType);
|
|
|
|
mSrcPrepCount = prepcount + readable;
|
|
NumSrcSamples = 0;
|
|
break;
|
|
}
|
|
|
|
const auto SrcData = al::span<float>{mSrcSamples};
|
|
const auto DstData = al::span<float>{mDstSamples};
|
|
uint DataPosFrac{mFracOffset};
|
|
uint64_t DataSize64{prepcount};
|
|
DataSize64 += readable;
|
|
DataSize64 -= MaxResamplerPadding;
|
|
DataSize64 <<= MixerFracBits;
|
|
DataSize64 -= DataPosFrac;
|
|
|
|
/* If we have a full prep, we can generate at least one sample. */
|
|
auto DstSize = static_cast<uint>(std::clamp((DataSize64 + increment-1)/increment, 1_u64,
|
|
uint64_t{BufferLineSize}));
|
|
DstSize = std::min(DstSize, dstframes-pos);
|
|
|
|
const uint DataPosEnd{DstSize*increment + DataPosFrac};
|
|
const uint SrcDataEnd{DataPosEnd>>MixerFracBits};
|
|
|
|
assert(prepcount+readable >= SrcDataEnd);
|
|
const uint nextprep{std::min(prepcount+readable-SrcDataEnd, MaxResamplerPadding)};
|
|
|
|
for(size_t chan{0u};chan < mChan.size();chan++)
|
|
{
|
|
/* Load the previous samples into the source data first, then the
|
|
* new samples from the input buffer.
|
|
*/
|
|
std::copy_n(mChan[chan].PrevSamples.cbegin(), prepcount, SrcData.begin());
|
|
LoadSamples(SrcData.subspan(prepcount, readable), SamplesIn.data(), chan, mChan.size(),
|
|
mSrcType);
|
|
|
|
/* Store as many prep samples for next time as possible, given the
|
|
* number of output samples being generated.
|
|
*/
|
|
auto previter = std::copy_n(SrcData.begin()+ptrdiff_t(SrcDataEnd), nextprep,
|
|
mChan[chan].PrevSamples.begin());
|
|
std::fill(previter, mChan[chan].PrevSamples.end(), 0.0f);
|
|
|
|
/* Now resample, and store the result in the output buffer. */
|
|
mResample(&mState, SrcData, DataPosFrac, increment, DstData.first(DstSize));
|
|
|
|
StoreSamples(SamplesOut.data(), DstData.first(DstSize), chan, mChan.size(), mDstType);
|
|
}
|
|
|
|
/* Update the number of prep samples still available, as well as the
|
|
* fractional offset.
|
|
*/
|
|
mSrcPrepCount = nextprep;
|
|
mFracOffset = DataPosEnd & MixerFracMask;
|
|
|
|
/* Update the src and dst pointers in case there's still more to do. */
|
|
const uint srcread{std::min(NumSrcSamples, SrcDataEnd + mSrcPrepCount - prepcount)};
|
|
SamplesIn = SamplesIn.subspan(SrcFrameSize*srcread);
|
|
NumSrcSamples -= srcread;
|
|
|
|
SamplesOut = SamplesOut.subspan(DstFrameSize*DstSize);
|
|
pos += DstSize;
|
|
}
|
|
|
|
*src = SamplesIn.data();
|
|
*srcframes = NumSrcSamples;
|
|
|
|
return pos;
|
|
}
|
|
|
|
uint SampleConverter::convertPlanar(const void **src, uint *srcframes, void *const*dst, uint dstframes)
|
|
{
|
|
const auto srcs = al::span{src, mChan.size()};
|
|
const auto dsts = al::span{dst, mChan.size()};
|
|
const uint increment{mIncrement};
|
|
uint NumSrcSamples{*srcframes};
|
|
|
|
FPUCtl mixer_mode{};
|
|
uint pos{0};
|
|
while(pos < dstframes && NumSrcSamples > 0)
|
|
{
|
|
const uint prepcount{mSrcPrepCount};
|
|
const uint readable{std::min(NumSrcSamples, uint{BufferLineSize} - prepcount)};
|
|
|
|
if(prepcount < MaxResamplerPadding && MaxResamplerPadding-prepcount >= readable)
|
|
{
|
|
/* Not enough input samples to generate an output sample. Store
|
|
* what we're given for later.
|
|
*/
|
|
for(size_t chan{0u};chan < mChan.size();chan++)
|
|
{
|
|
auto samples = al::span{static_cast<const std::byte*>(srcs[chan]),
|
|
NumSrcSamples*size_t{mSrcTypeSize}};
|
|
LoadSamples(al::span{mChan[chan].PrevSamples}.subspan(prepcount, readable),
|
|
samples.data(), 0, 1, mSrcType);
|
|
srcs[chan] = samples.subspan(size_t{mSrcTypeSize}*readable).data();
|
|
}
|
|
|
|
mSrcPrepCount = prepcount + readable;
|
|
NumSrcSamples = 0;
|
|
break;
|
|
}
|
|
|
|
const auto SrcData = al::span{mSrcSamples};
|
|
const auto DstData = al::span{mDstSamples};
|
|
uint DataPosFrac{mFracOffset};
|
|
uint64_t DataSize64{prepcount};
|
|
DataSize64 += readable;
|
|
DataSize64 -= MaxResamplerPadding;
|
|
DataSize64 <<= MixerFracBits;
|
|
DataSize64 -= DataPosFrac;
|
|
|
|
/* If we have a full prep, we can generate at least one sample. */
|
|
auto DstSize = static_cast<uint>(std::clamp((DataSize64 + increment-1)/increment, 1_u64,
|
|
uint64_t{BufferLineSize}));
|
|
DstSize = std::min(DstSize, dstframes-pos);
|
|
|
|
const uint DataPosEnd{DstSize*increment + DataPosFrac};
|
|
const uint SrcDataEnd{DataPosEnd>>MixerFracBits};
|
|
|
|
assert(prepcount+readable >= SrcDataEnd);
|
|
const uint nextprep{std::min(prepcount+readable-SrcDataEnd, MaxResamplerPadding)};
|
|
|
|
for(size_t chan{0u};chan < mChan.size();chan++)
|
|
{
|
|
/* Load the previous samples into the source data first, then the
|
|
* new samples from the input buffer.
|
|
*/
|
|
auto srciter = std::copy_n(mChan[chan].PrevSamples.cbegin(),prepcount,SrcData.begin());
|
|
LoadSamples({srciter, readable}, srcs[chan], 0, 1, mSrcType);
|
|
|
|
/* Store as many prep samples for next time as possible, given the
|
|
* number of output samples being generated.
|
|
*/
|
|
auto previter = std::copy_n(SrcData.begin()+ptrdiff_t(SrcDataEnd), nextprep,
|
|
mChan[chan].PrevSamples.begin());
|
|
std::fill(previter, mChan[chan].PrevSamples.end(), 0.0f);
|
|
|
|
/* Now resample, and store the result in the output buffer. */
|
|
mResample(&mState, SrcData, DataPosFrac, increment, DstData.first(DstSize));
|
|
|
|
auto DstSamples = al::span{static_cast<std::byte*>(dsts[chan]),
|
|
size_t{mDstTypeSize}*dstframes}.subspan(pos*size_t{mDstTypeSize});
|
|
StoreSamples(DstSamples.data(), DstData.first(DstSize), 0, 1, mDstType);
|
|
}
|
|
|
|
/* Update the number of prep samples still available, as well as the
|
|
* fractional offset.
|
|
*/
|
|
mSrcPrepCount = nextprep;
|
|
mFracOffset = DataPosEnd & MixerFracMask;
|
|
|
|
/* Update the src and dst pointers in case there's still more to do. */
|
|
const uint srcread{std::min(NumSrcSamples, SrcDataEnd + mSrcPrepCount - prepcount)};
|
|
std::for_each(srcs.begin(), srcs.end(), [this,NumSrcSamples,srcread](const void *&srcref)
|
|
{
|
|
auto srcspan = al::span{static_cast<const std::byte*>(srcref),
|
|
size_t{mSrcTypeSize}*NumSrcSamples};
|
|
srcref = srcspan.subspan(size_t{mSrcTypeSize}*srcread).data();
|
|
});
|
|
NumSrcSamples -= srcread;
|
|
|
|
pos += DstSize;
|
|
}
|
|
|
|
*srcframes = NumSrcSamples;
|
|
|
|
return pos;
|
|
}
|
|
|
|
|
|
void ChannelConverter::convert(const void *src, float *dst, uint frames) const
|
|
{
|
|
if(mDstChans == DevFmtMono)
|
|
{
|
|
const float scale{std::sqrt(1.0f / static_cast<float>(al::popcount(mChanMask)))};
|
|
switch(mSrcType)
|
|
{
|
|
#define HANDLE_FMT(T) case T: Multi2Mono<T>(mChanMask, mSrcStep, scale, {dst, frames}, src); break
|
|
HANDLE_FMT(DevFmtByte);
|
|
HANDLE_FMT(DevFmtUByte);
|
|
HANDLE_FMT(DevFmtShort);
|
|
HANDLE_FMT(DevFmtUShort);
|
|
HANDLE_FMT(DevFmtInt);
|
|
HANDLE_FMT(DevFmtUInt);
|
|
HANDLE_FMT(DevFmtFloat);
|
|
#undef HANDLE_FMT
|
|
}
|
|
}
|
|
else if(mChanMask == 0x1 && mDstChans == DevFmtStereo)
|
|
{
|
|
switch(mSrcType)
|
|
{
|
|
#define HANDLE_FMT(T) case T: Mono2Stereo<T>({dst, frames*2_uz}, src); break
|
|
HANDLE_FMT(DevFmtByte);
|
|
HANDLE_FMT(DevFmtUByte);
|
|
HANDLE_FMT(DevFmtShort);
|
|
HANDLE_FMT(DevFmtUShort);
|
|
HANDLE_FMT(DevFmtInt);
|
|
HANDLE_FMT(DevFmtUInt);
|
|
HANDLE_FMT(DevFmtFloat);
|
|
#undef HANDLE_FMT
|
|
}
|
|
}
|
|
}
|