converter.cpp 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457
  1. #include "config.h"
  2. #include "converter.h"
  3. #include <algorithm>
  4. #include <cassert>
  5. #include <cmath>
  6. #include <cstddef>
  7. #include <cstdint>
  8. #include <iterator>
  9. #include <climits>
  10. #include "albit.h"
  11. #include "alnumeric.h"
  12. #include "fpu_ctrl.h"
  13. namespace {
  14. constexpr uint MaxPitch{10};
  15. static_assert((BufferLineSize-1)/MaxPitch > 0, "MaxPitch is too large for BufferLineSize!");
  16. static_assert((INT_MAX>>MixerFracBits)/MaxPitch > BufferLineSize,
  17. "MaxPitch and/or BufferLineSize are too large for MixerFracBits!");
  18. template<DevFmtType T>
  19. constexpr float LoadSample(DevFmtType_t<T> val) noexcept = delete;
  20. template<> constexpr float LoadSample<DevFmtByte>(DevFmtType_t<DevFmtByte> val) noexcept
  21. { return float(val) * (1.0f/128.0f); }
  22. template<> constexpr float LoadSample<DevFmtShort>(DevFmtType_t<DevFmtShort> val) noexcept
  23. { return float(val) * (1.0f/32768.0f); }
  24. template<> constexpr float LoadSample<DevFmtInt>(DevFmtType_t<DevFmtInt> val) noexcept
  25. { return static_cast<float>(val) * (1.0f/2147483648.0f); }
  26. template<> constexpr float LoadSample<DevFmtFloat>(DevFmtType_t<DevFmtFloat> val) noexcept
  27. { return val; }
  28. template<> constexpr float LoadSample<DevFmtUByte>(DevFmtType_t<DevFmtUByte> val) noexcept
  29. { return LoadSample<DevFmtByte>(static_cast<int8_t>(val - 128)); }
  30. template<> constexpr float LoadSample<DevFmtUShort>(DevFmtType_t<DevFmtUShort> val) noexcept
  31. { return LoadSample<DevFmtShort>(static_cast<int16_t>(val - 32768)); }
  32. template<> constexpr float LoadSample<DevFmtUInt>(DevFmtType_t<DevFmtUInt> val) noexcept
  33. { return LoadSample<DevFmtInt>(static_cast<int32_t>(val - 2147483648u)); }
  34. template<DevFmtType T>
  35. inline void LoadSampleArray(const al::span<float> dst, const void *src, const size_t channel,
  36. const size_t srcstep) noexcept
  37. {
  38. assert(channel < srcstep);
  39. const auto srcspan = al::span{static_cast<const DevFmtType_t<T>*>(src), dst.size()*srcstep};
  40. auto ssrc = srcspan.cbegin();
  41. std::generate(dst.begin(), dst.end(), [&ssrc,channel,srcstep]
  42. {
  43. const float ret{LoadSample<T>(ssrc[channel])};
  44. ssrc += ptrdiff_t(srcstep);
  45. return ret;
  46. });
  47. }
  48. void LoadSamples(const al::span<float> dst, const void *src, const size_t channel,
  49. const size_t srcstep, const DevFmtType srctype) noexcept
  50. {
  51. #define HANDLE_FMT(T) \
  52. case T: LoadSampleArray<T>(dst, src, channel, srcstep); break
  53. switch(srctype)
  54. {
  55. HANDLE_FMT(DevFmtByte);
  56. HANDLE_FMT(DevFmtUByte);
  57. HANDLE_FMT(DevFmtShort);
  58. HANDLE_FMT(DevFmtUShort);
  59. HANDLE_FMT(DevFmtInt);
  60. HANDLE_FMT(DevFmtUInt);
  61. HANDLE_FMT(DevFmtFloat);
  62. }
  63. #undef HANDLE_FMT
  64. }
  65. template<DevFmtType T>
  66. inline DevFmtType_t<T> StoreSample(float) noexcept;
  67. template<> inline float StoreSample<DevFmtFloat>(float val) noexcept
  68. { return val; }
  69. template<> inline int32_t StoreSample<DevFmtInt>(float val) noexcept
  70. { return fastf2i(std::clamp(val*2147483648.0f, -2147483648.0f, 2147483520.0f)); }
  71. template<> inline int16_t StoreSample<DevFmtShort>(float val) noexcept
  72. { return static_cast<int16_t>(fastf2i(std::clamp(val*32768.0f, -32768.0f, 32767.0f))); }
  73. template<> inline int8_t StoreSample<DevFmtByte>(float val) noexcept
  74. { return static_cast<int8_t>(fastf2i(std::clamp(val*128.0f, -128.0f, 127.0f))); }
  75. /* Define unsigned output variations. */
  76. template<> inline uint32_t StoreSample<DevFmtUInt>(float val) noexcept
  77. { return static_cast<uint32_t>(StoreSample<DevFmtInt>(val)) + 2147483648u; }
  78. template<> inline uint16_t StoreSample<DevFmtUShort>(float val) noexcept
  79. { return static_cast<uint16_t>(StoreSample<DevFmtShort>(val) + 32768); }
  80. template<> inline uint8_t StoreSample<DevFmtUByte>(float val) noexcept
  81. { return static_cast<uint8_t>(StoreSample<DevFmtByte>(val) + 128); }
  82. template<DevFmtType T>
  83. inline void StoreSampleArray(void *dst, const al::span<const float> src, const size_t channel,
  84. const size_t dststep) noexcept
  85. {
  86. assert(channel < dststep);
  87. const auto dstspan = al::span{static_cast<DevFmtType_t<T>*>(dst), src.size()*dststep};
  88. auto sdst = dstspan.begin();
  89. std::for_each(src.cbegin(), src.cend(), [&sdst,channel,dststep](const float in)
  90. {
  91. sdst[channel] = StoreSample<T>(in);
  92. sdst += ptrdiff_t(dststep);
  93. });
  94. }
  95. void StoreSamples(void *dst, const al::span<const float> src, const size_t channel,
  96. const size_t dststep, const DevFmtType dsttype) noexcept
  97. {
  98. #define HANDLE_FMT(T) \
  99. case T: StoreSampleArray<T>(dst, src, channel, dststep); break
  100. switch(dsttype)
  101. {
  102. HANDLE_FMT(DevFmtByte);
  103. HANDLE_FMT(DevFmtUByte);
  104. HANDLE_FMT(DevFmtShort);
  105. HANDLE_FMT(DevFmtUShort);
  106. HANDLE_FMT(DevFmtInt);
  107. HANDLE_FMT(DevFmtUInt);
  108. HANDLE_FMT(DevFmtFloat);
  109. }
  110. #undef HANDLE_FMT
  111. }
  112. template<DevFmtType T>
  113. void Mono2Stereo(const al::span<float> dst, const void *src) noexcept
  114. {
  115. const auto srcspan = al::span{static_cast<const DevFmtType_t<T>*>(src), dst.size()>>1};
  116. auto sdst = dst.begin();
  117. std::for_each(srcspan.cbegin(), srcspan.cend(), [&sdst](const auto in)
  118. { sdst = std::fill_n(sdst, 2, LoadSample<T>(in)*0.707106781187f); });
  119. }
  120. template<DevFmtType T>
  121. void Multi2Mono(uint chanmask, const size_t step, const float scale, const al::span<float> dst,
  122. const void *src) noexcept
  123. {
  124. const auto srcspan = al::span{static_cast<const DevFmtType_t<T>*>(src), step*dst.size()};
  125. std::fill_n(dst.begin(), dst.size(), 0.0f);
  126. for(size_t c{0};chanmask;++c)
  127. {
  128. if((chanmask&1)) LIKELY
  129. {
  130. auto ssrc = srcspan.cbegin();
  131. std::for_each(dst.begin(), dst.end(), [&ssrc,step,c](float &sample)
  132. {
  133. const float s{LoadSample<T>(ssrc[c])};
  134. ssrc += ptrdiff_t(step);
  135. sample += s;
  136. });
  137. }
  138. chanmask >>= 1;
  139. }
  140. std::for_each(dst.begin(), dst.end(), [scale](float &sample) noexcept { sample *= scale; });
  141. }
  142. } // namespace
  143. SampleConverterPtr SampleConverter::Create(DevFmtType srcType, DevFmtType dstType, size_t numchans,
  144. uint srcRate, uint dstRate, Resampler resampler)
  145. {
  146. if(numchans < 1 || srcRate < 1 || dstRate < 1)
  147. return nullptr;
  148. SampleConverterPtr converter{new(FamCount(numchans)) SampleConverter{numchans}};
  149. converter->mSrcType = srcType;
  150. converter->mDstType = dstType;
  151. converter->mSrcTypeSize = BytesFromDevFmt(srcType);
  152. converter->mDstTypeSize = BytesFromDevFmt(dstType);
  153. converter->mSrcPrepCount = MaxResamplerPadding;
  154. converter->mFracOffset = 0;
  155. for(auto &chan : converter->mChan)
  156. chan.PrevSamples.fill(0.0f);
  157. /* Have to set the mixer FPU mode since that's what the resampler code expects. */
  158. FPUCtl mixer_mode{};
  159. const auto step = std::min(std::round(srcRate*double{MixerFracOne}/dstRate),
  160. MaxPitch*double{MixerFracOne});
  161. converter->mIncrement = std::max(static_cast<uint>(step), 1u);
  162. if(converter->mIncrement == MixerFracOne)
  163. {
  164. converter->mResample = [](const InterpState*, const al::span<const float> src, uint,
  165. const uint, const al::span<float> dst)
  166. { std::copy_n(src.begin()+MaxResamplerEdge, dst.size(), dst.begin()); };
  167. }
  168. else
  169. converter->mResample = PrepareResampler(resampler, converter->mIncrement,
  170. &converter->mState);
  171. return converter;
  172. }
  173. uint SampleConverter::availableOut(uint srcframes) const
  174. {
  175. if(srcframes < 1)
  176. {
  177. /* No output samples if there's no input samples. */
  178. return 0;
  179. }
  180. const uint prepcount{mSrcPrepCount};
  181. if(prepcount < MaxResamplerPadding && MaxResamplerPadding - prepcount >= srcframes)
  182. {
  183. /* Not enough input samples to generate an output sample. */
  184. return 0;
  185. }
  186. uint64_t DataSize64{prepcount};
  187. DataSize64 += srcframes;
  188. DataSize64 -= MaxResamplerPadding;
  189. DataSize64 <<= MixerFracBits;
  190. DataSize64 -= mFracOffset;
  191. /* If we have a full prep, we can generate at least one sample. */
  192. return static_cast<uint>(std::clamp((DataSize64 + mIncrement-1)/mIncrement, 1_u64,
  193. uint64_t{std::numeric_limits<int>::max()}));
  194. }
  195. uint SampleConverter::convert(const void **src, uint *srcframes, void *dst, uint dstframes)
  196. {
  197. const size_t SrcFrameSize{mChan.size() * mSrcTypeSize};
  198. const size_t DstFrameSize{mChan.size() * mDstTypeSize};
  199. const uint increment{mIncrement};
  200. uint NumSrcSamples{*srcframes};
  201. auto SamplesIn = al::span{static_cast<const std::byte*>(*src), NumSrcSamples*SrcFrameSize};
  202. auto SamplesOut = al::span{static_cast<std::byte*>(dst), dstframes*DstFrameSize};
  203. FPUCtl mixer_mode{};
  204. uint pos{0};
  205. while(pos < dstframes && NumSrcSamples > 0)
  206. {
  207. const uint prepcount{mSrcPrepCount};
  208. const uint readable{std::min(NumSrcSamples, uint{BufferLineSize} - prepcount)};
  209. if(prepcount < MaxResamplerPadding && MaxResamplerPadding-prepcount >= readable)
  210. {
  211. /* Not enough input samples to generate an output sample. Store
  212. * what we're given for later.
  213. */
  214. for(size_t chan{0u};chan < mChan.size();chan++)
  215. LoadSamples(al::span{mChan[chan].PrevSamples}.subspan(prepcount, readable),
  216. SamplesIn.data(), chan, mChan.size(), mSrcType);
  217. mSrcPrepCount = prepcount + readable;
  218. NumSrcSamples = 0;
  219. break;
  220. }
  221. const auto SrcData = al::span<float>{mSrcSamples};
  222. const auto DstData = al::span<float>{mDstSamples};
  223. uint DataPosFrac{mFracOffset};
  224. uint64_t DataSize64{prepcount};
  225. DataSize64 += readable;
  226. DataSize64 -= MaxResamplerPadding;
  227. DataSize64 <<= MixerFracBits;
  228. DataSize64 -= DataPosFrac;
  229. /* If we have a full prep, we can generate at least one sample. */
  230. auto DstSize = static_cast<uint>(std::clamp((DataSize64 + increment-1)/increment, 1_u64,
  231. uint64_t{BufferLineSize}));
  232. DstSize = std::min(DstSize, dstframes-pos);
  233. const uint DataPosEnd{DstSize*increment + DataPosFrac};
  234. const uint SrcDataEnd{DataPosEnd>>MixerFracBits};
  235. assert(prepcount+readable >= SrcDataEnd);
  236. const uint nextprep{std::min(prepcount+readable-SrcDataEnd, MaxResamplerPadding)};
  237. for(size_t chan{0u};chan < mChan.size();chan++)
  238. {
  239. /* Load the previous samples into the source data first, then the
  240. * new samples from the input buffer.
  241. */
  242. std::copy_n(mChan[chan].PrevSamples.cbegin(), prepcount, SrcData.begin());
  243. LoadSamples(SrcData.subspan(prepcount, readable), SamplesIn.data(), chan, mChan.size(),
  244. mSrcType);
  245. /* Store as many prep samples for next time as possible, given the
  246. * number of output samples being generated.
  247. */
  248. auto previter = std::copy_n(SrcData.begin()+ptrdiff_t(SrcDataEnd), nextprep,
  249. mChan[chan].PrevSamples.begin());
  250. std::fill(previter, mChan[chan].PrevSamples.end(), 0.0f);
  251. /* Now resample, and store the result in the output buffer. */
  252. mResample(&mState, SrcData, DataPosFrac, increment, DstData.first(DstSize));
  253. StoreSamples(SamplesOut.data(), DstData.first(DstSize), chan, mChan.size(), mDstType);
  254. }
  255. /* Update the number of prep samples still available, as well as the
  256. * fractional offset.
  257. */
  258. mSrcPrepCount = nextprep;
  259. mFracOffset = DataPosEnd & MixerFracMask;
  260. /* Update the src and dst pointers in case there's still more to do. */
  261. const uint srcread{std::min(NumSrcSamples, SrcDataEnd + mSrcPrepCount - prepcount)};
  262. SamplesIn = SamplesIn.subspan(SrcFrameSize*srcread);
  263. NumSrcSamples -= srcread;
  264. SamplesOut = SamplesOut.subspan(DstFrameSize*DstSize);
  265. pos += DstSize;
  266. }
  267. *src = SamplesIn.data();
  268. *srcframes = NumSrcSamples;
  269. return pos;
  270. }
  271. uint SampleConverter::convertPlanar(const void **src, uint *srcframes, void *const*dst, uint dstframes)
  272. {
  273. const auto srcs = al::span{src, mChan.size()};
  274. const auto dsts = al::span{dst, mChan.size()};
  275. const uint increment{mIncrement};
  276. uint NumSrcSamples{*srcframes};
  277. FPUCtl mixer_mode{};
  278. uint pos{0};
  279. while(pos < dstframes && NumSrcSamples > 0)
  280. {
  281. const uint prepcount{mSrcPrepCount};
  282. const uint readable{std::min(NumSrcSamples, uint{BufferLineSize} - prepcount)};
  283. if(prepcount < MaxResamplerPadding && MaxResamplerPadding-prepcount >= readable)
  284. {
  285. /* Not enough input samples to generate an output sample. Store
  286. * what we're given for later.
  287. */
  288. for(size_t chan{0u};chan < mChan.size();chan++)
  289. {
  290. auto samples = al::span{static_cast<const std::byte*>(srcs[chan]),
  291. NumSrcSamples*size_t{mSrcTypeSize}};
  292. LoadSamples(al::span{mChan[chan].PrevSamples}.subspan(prepcount, readable),
  293. samples.data(), 0, 1, mSrcType);
  294. srcs[chan] = samples.subspan(size_t{mSrcTypeSize}*readable).data();
  295. }
  296. mSrcPrepCount = prepcount + readable;
  297. NumSrcSamples = 0;
  298. break;
  299. }
  300. const auto SrcData = al::span{mSrcSamples};
  301. const auto DstData = al::span{mDstSamples};
  302. uint DataPosFrac{mFracOffset};
  303. uint64_t DataSize64{prepcount};
  304. DataSize64 += readable;
  305. DataSize64 -= MaxResamplerPadding;
  306. DataSize64 <<= MixerFracBits;
  307. DataSize64 -= DataPosFrac;
  308. /* If we have a full prep, we can generate at least one sample. */
  309. auto DstSize = static_cast<uint>(std::clamp((DataSize64 + increment-1)/increment, 1_u64,
  310. uint64_t{BufferLineSize}));
  311. DstSize = std::min(DstSize, dstframes-pos);
  312. const uint DataPosEnd{DstSize*increment + DataPosFrac};
  313. const uint SrcDataEnd{DataPosEnd>>MixerFracBits};
  314. assert(prepcount+readable >= SrcDataEnd);
  315. const uint nextprep{std::min(prepcount+readable-SrcDataEnd, MaxResamplerPadding)};
  316. for(size_t chan{0u};chan < mChan.size();chan++)
  317. {
  318. /* Load the previous samples into the source data first, then the
  319. * new samples from the input buffer.
  320. */
  321. auto srciter = std::copy_n(mChan[chan].PrevSamples.cbegin(),prepcount,SrcData.begin());
  322. LoadSamples({srciter, readable}, srcs[chan], 0, 1, mSrcType);
  323. /* Store as many prep samples for next time as possible, given the
  324. * number of output samples being generated.
  325. */
  326. auto previter = std::copy_n(SrcData.begin()+ptrdiff_t(SrcDataEnd), nextprep,
  327. mChan[chan].PrevSamples.begin());
  328. std::fill(previter, mChan[chan].PrevSamples.end(), 0.0f);
  329. /* Now resample, and store the result in the output buffer. */
  330. mResample(&mState, SrcData, DataPosFrac, increment, DstData.first(DstSize));
  331. auto DstSamples = al::span{static_cast<std::byte*>(dsts[chan]),
  332. size_t{mDstTypeSize}*dstframes}.subspan(pos*size_t{mDstTypeSize});
  333. StoreSamples(DstSamples.data(), DstData.first(DstSize), 0, 1, mDstType);
  334. }
  335. /* Update the number of prep samples still available, as well as the
  336. * fractional offset.
  337. */
  338. mSrcPrepCount = nextprep;
  339. mFracOffset = DataPosEnd & MixerFracMask;
  340. /* Update the src and dst pointers in case there's still more to do. */
  341. const uint srcread{std::min(NumSrcSamples, SrcDataEnd + mSrcPrepCount - prepcount)};
  342. std::for_each(srcs.begin(), srcs.end(), [this,NumSrcSamples,srcread](const void *&srcref)
  343. {
  344. auto srcspan = al::span{static_cast<const std::byte*>(srcref),
  345. size_t{mSrcTypeSize}*NumSrcSamples};
  346. srcref = srcspan.subspan(size_t{mSrcTypeSize}*srcread).data();
  347. });
  348. NumSrcSamples -= srcread;
  349. pos += DstSize;
  350. }
  351. *srcframes = NumSrcSamples;
  352. return pos;
  353. }
  354. void ChannelConverter::convert(const void *src, float *dst, uint frames) const
  355. {
  356. if(mDstChans == DevFmtMono)
  357. {
  358. const float scale{std::sqrt(1.0f / static_cast<float>(al::popcount(mChanMask)))};
  359. switch(mSrcType)
  360. {
  361. #define HANDLE_FMT(T) case T: Multi2Mono<T>(mChanMask, mSrcStep, scale, {dst, frames}, src); break
  362. HANDLE_FMT(DevFmtByte);
  363. HANDLE_FMT(DevFmtUByte);
  364. HANDLE_FMT(DevFmtShort);
  365. HANDLE_FMT(DevFmtUShort);
  366. HANDLE_FMT(DevFmtInt);
  367. HANDLE_FMT(DevFmtUInt);
  368. HANDLE_FMT(DevFmtFloat);
  369. #undef HANDLE_FMT
  370. }
  371. }
  372. else if(mChanMask == 0x1 && mDstChans == DevFmtStereo)
  373. {
  374. switch(mSrcType)
  375. {
  376. #define HANDLE_FMT(T) case T: Mono2Stereo<T>({dst, frames*2_uz}, src); break
  377. HANDLE_FMT(DevFmtByte);
  378. HANDLE_FMT(DevFmtUByte);
  379. HANDLE_FMT(DevFmtShort);
  380. HANDLE_FMT(DevFmtUShort);
  381. HANDLE_FMT(DevFmtInt);
  382. HANDLE_FMT(DevFmtUInt);
  383. HANDLE_FMT(DevFmtFloat);
  384. #undef HANDLE_FMT
  385. }
  386. }
  387. }