1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108
|
/*
* Copyright (C) 2020 Linux Studio Plugins Project <https://lsp-plug.in/>
* (C) 2020 Vladimir Sadovnikov <sadko4u@gmail.com>
*
* This file is part of lsp-dsp-lib
* Created on: 31 мар. 2020 г.
*
* lsp-dsp-lib is free software: you can redistribute it and/or modify
* it under the terms of the GNU Lesser General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* any later version.
*
* lsp-dsp-lib is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public License
* along with lsp-dsp-lib. If not, see <https://www.gnu.org/licenses/>.
*/
#ifndef PRIVATE_DSP_ARCH_AARCH64_ASIMD_FFT_NORMALIZE_H_
#define PRIVATE_DSP_ARCH_AARCH64_ASIMD_FFT_NORMALIZE_H_
#ifndef PRIVATE_DSP_ARCH_AARCH64_ASIMD_IMPL
#error "This header should not be included directly"
#endif /* PRIVATE_DSP_ARCH_AARCH64_ASIMD_IMPL */
namespace lsp
{
namespace asimd
{
void normalize_fft3(float *dre, float *dim, const float *re, const float *im, size_t rank)
{
IF_ARCH_AARCH64(
float k = 1.0f/(1 << rank);
size_t count = 1 << rank;
);
ARCH_AARCH64_ASM(
// x8 blocks
__ASM_EMIT("dup v0.4s, %[k].s[0]") // v0 = k
__ASM_EMIT("subs %[count], %[count], #8")
__ASM_EMIT("mov v1.16b, v0.16b")
__ASM_EMIT("b.lt 2f")
__ASM_EMIT("1:")
__ASM_EMIT("ldp q4, q5, [%[s_re], #0x00]")
__ASM_EMIT("ldp q6, q7, [%[s_im], #0x00]")
__ASM_EMIT("fmul v4.4s, v4.4s, v0.4s")
__ASM_EMIT("fmul v5.4s, v5.4s, v1.4s")
__ASM_EMIT("fmul v6.4s, v6.4s, v0.4s")
__ASM_EMIT("fmul v7.4s, v7.4s, v1.4s")
__ASM_EMIT("stp q4, q5, [%[d_re], #0x00]")
__ASM_EMIT("stp q6, q7, [%[d_im], #0x00]")
__ASM_EMIT("subs %[count], %[count], #8")
__ASM_EMIT("add %[s_re], %[s_re], #0x20")
__ASM_EMIT("add %[s_im], %[s_im], #0x20")
__ASM_EMIT("add %[d_re], %[d_re], #0x20")
__ASM_EMIT("add %[d_im], %[d_im], #0x20")
__ASM_EMIT("b.hs 1b")
__ASM_EMIT("2:")
: [s_re] "+r" (re), [s_im] "+r" (im),
[d_re] "+r" (dre), [d_im] "+r" (dim),
[k] "+w" (k), [count] "+r" (count)
:
: "cc", "memory",
"v0", "v1",
"v4", "v5", "v6", "v7"
);
}
void normalize_fft2(float *re, float *im, size_t rank)
{
IF_ARCH_AARCH64(
float k = 1.0f/(1 << rank);
size_t count = 1 << rank;
);
ARCH_AARCH64_ASM(
// x8 blocks
__ASM_EMIT("dup v0.4s, %[k].s[0]") // v0 = k
__ASM_EMIT("subs %[count], %[count], #8")
__ASM_EMIT("mov v1.16b, v0.16b")
__ASM_EMIT("b.lt 2f")
__ASM_EMIT("1:")
__ASM_EMIT("ldp q4, q5, [%[d_re], #0x00]")
__ASM_EMIT("ldp q6, q7, [%[d_im], #0x00]")
__ASM_EMIT("fmul v4.4s, v4.4s, v0.4s")
__ASM_EMIT("fmul v5.4s, v5.4s, v1.4s")
__ASM_EMIT("fmul v6.4s, v6.4s, v0.4s")
__ASM_EMIT("fmul v7.4s, v7.4s, v1.4s")
__ASM_EMIT("stp q4, q5, [%[d_re], #0x00]")
__ASM_EMIT("stp q6, q7, [%[d_im], #0x00]")
__ASM_EMIT("subs %[count], %[count], #8")
__ASM_EMIT("add %[d_re], %[d_re], #0x20")
__ASM_EMIT("add %[d_im], %[d_im], #0x20")
__ASM_EMIT("b.hs 1b")
__ASM_EMIT("2:")
: [d_re] "+r" (re), [d_im] "+r" (im),
[k] "+w" (k), [count] "+r" (count)
:
: "cc", "memory",
"v0", "v1",
"v4", "v5", "v6", "v7"
);
}
}
}
#endif /* PRIVATE_DSP_ARCH_AARCH64_ASIMD_FFT_NORMALIZE_H_ */
|