arm_cfft_f16.c 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680
  1. /* ----------------------------------------------------------------------
  2. * Project: CMSIS DSP Library
  3. * Title: arm_cfft_f32.c
  4. * Description: Combined Radix Decimation in Frequency CFFT Floating point processing function
  5. *
  6. * $Date: 23 April 2021
  7. * $Revision: V1.9.0
  8. *
  9. * Target Processor: Cortex-M and Cortex-A cores
  10. * -------------------------------------------------------------------- */
  11. /*
  12. * Copyright (C) 2010-2021 ARM Limited or its affiliates. All rights reserved.
  13. *
  14. * SPDX-License-Identifier: Apache-2.0
  15. *
  16. * Licensed under the Apache License, Version 2.0 (the License); you may
  17. * not use this file except in compliance with the License.
  18. * You may obtain a copy of the License at
  19. *
  20. * www.apache.org/licenses/LICENSE-2.0
  21. *
  22. * Unless required by applicable law or agreed to in writing, software
  23. * distributed under the License is distributed on an AS IS BASIS, WITHOUT
  24. * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  25. * See the License for the specific language governing permissions and
  26. * limitations under the License.
  27. */
  28. #include "dsp/transform_functions_f16.h"
  29. #include "arm_common_tables_f16.h"
  30. #if defined(ARM_MATH_MVE_FLOAT16) && !defined(ARM_MATH_AUTOVECTORIZE)
  31. #include "arm_helium_utils.h"
  32. #include "arm_vec_fft.h"
  33. #include "arm_mve_tables_f16.h"
  34. static float16_t arm_inverse_fft_length_f16(uint16_t fftLen)
  35. {
  36. float16_t retValue=1.0;
  37. switch (fftLen)
  38. {
  39. case 4096U:
  40. retValue = (float16_t)0.000244140625f;
  41. break;
  42. case 2048U:
  43. retValue = (float16_t)0.00048828125f;
  44. break;
  45. case 1024U:
  46. retValue = (float16_t)0.0009765625f;
  47. break;
  48. case 512U:
  49. retValue = (float16_t)0.001953125f;
  50. break;
  51. case 256U:
  52. retValue = (float16_t)0.00390625f;
  53. break;
  54. case 128U:
  55. retValue = (float16_t)0.0078125f;
  56. break;
  57. case 64U:
  58. retValue = (float16_t)0.015625f;
  59. break;
  60. case 32U:
  61. retValue = (float16_t)0.03125f;
  62. break;
  63. case 16U:
  64. retValue = (float16_t)0.0625f;
  65. break;
  66. default:
  67. break;
  68. }
  69. return(retValue);
  70. }
  71. static void _arm_radix4_butterfly_f16_mve(const arm_cfft_instance_f16 * S,float16_t * pSrc, uint32_t fftLen)
  72. {
  73. f16x8_t vecTmp0, vecTmp1;
  74. f16x8_t vecSum0, vecDiff0, vecSum1, vecDiff1;
  75. f16x8_t vecA, vecB, vecC, vecD;
  76. uint32_t blkCnt;
  77. uint32_t n1, n2;
  78. uint32_t stage = 0;
  79. int32_t iter = 1;
  80. static const int32_t strides[4] =
  81. { ( 0 - 16) * (int32_t)sizeof(float16_t *)
  82. , ( 4 - 16) * (int32_t)sizeof(float16_t *)
  83. , ( 8 - 16) * (int32_t)sizeof(float16_t *)
  84. , (12 - 16) * (int32_t)sizeof(float16_t *)};
  85. n2 = fftLen;
  86. n1 = n2;
  87. n2 >>= 2u;
  88. for (int k = fftLen / 4u; k > 1; k >>= 2)
  89. {
  90. float16_t const *p_rearranged_twiddle_tab_stride1 =
  91. &S->rearranged_twiddle_stride1[
  92. S->rearranged_twiddle_tab_stride1_arr[stage]];
  93. float16_t const *p_rearranged_twiddle_tab_stride2 =
  94. &S->rearranged_twiddle_stride2[
  95. S->rearranged_twiddle_tab_stride2_arr[stage]];
  96. float16_t const *p_rearranged_twiddle_tab_stride3 =
  97. &S->rearranged_twiddle_stride3[
  98. S->rearranged_twiddle_tab_stride3_arr[stage]];
  99. float16_t * pBase = pSrc;
  100. for (int i = 0; i < iter; i++)
  101. {
  102. float16_t *inA = pBase;
  103. float16_t *inB = inA + n2 * CMPLX_DIM;
  104. float16_t *inC = inB + n2 * CMPLX_DIM;
  105. float16_t *inD = inC + n2 * CMPLX_DIM;
  106. float16_t const *pW1 = p_rearranged_twiddle_tab_stride1;
  107. float16_t const *pW2 = p_rearranged_twiddle_tab_stride2;
  108. float16_t const *pW3 = p_rearranged_twiddle_tab_stride3;
  109. f16x8_t vecW;
  110. blkCnt = n2 / 4;
  111. /*
  112. * load 2 f16 complex pair
  113. */
  114. vecA = vldrhq_f16(inA);
  115. vecC = vldrhq_f16(inC);
  116. while (blkCnt > 0U)
  117. {
  118. vecB = vldrhq_f16(inB);
  119. vecD = vldrhq_f16(inD);
  120. vecSum0 = vecA + vecC; /* vecSum0 = vaddq(vecA, vecC) */
  121. vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */
  122. vecSum1 = vecB + vecD;
  123. vecDiff1 = vecB - vecD;
  124. /*
  125. * [ 1 1 1 1 ] * [ A B C D ]' .* 1
  126. */
  127. vecTmp0 = vecSum0 + vecSum1;
  128. vst1q(inA, vecTmp0);
  129. inA += 8;
  130. /*
  131. * [ 1 -1 1 -1 ] * [ A B C D ]'
  132. */
  133. vecTmp0 = vecSum0 - vecSum1;
  134. /*
  135. * [ 1 -1 1 -1 ] * [ A B C D ]'.* W2
  136. */
  137. vecW = vld1q(pW2);
  138. pW2 += 8;
  139. vecTmp1 = MVE_CMPLX_MULT_FLT_Conj_AxB(vecW, vecTmp0);
  140. vst1q(inB, vecTmp1);
  141. inB += 8;
  142. /*
  143. * [ 1 -i -1 +i ] * [ A B C D ]'
  144. */
  145. vecTmp0 = MVE_CMPLX_SUB_A_ixB(vecDiff0, vecDiff1);
  146. /*
  147. * [ 1 -i -1 +i ] * [ A B C D ]'.* W1
  148. */
  149. vecW = vld1q(pW1);
  150. pW1 +=8;
  151. vecTmp1 = MVE_CMPLX_MULT_FLT_Conj_AxB(vecW, vecTmp0);
  152. vst1q(inC, vecTmp1);
  153. inC += 8;
  154. /*
  155. * [ 1 +i -1 -i ] * [ A B C D ]'
  156. */
  157. vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1);
  158. /*
  159. * [ 1 +i -1 -i ] * [ A B C D ]'.* W3
  160. */
  161. vecW = vld1q(pW3);
  162. pW3 += 8;
  163. vecTmp1 = MVE_CMPLX_MULT_FLT_Conj_AxB(vecW, vecTmp0);
  164. vst1q(inD, vecTmp1);
  165. inD += 8;
  166. vecA = vldrhq_f16(inA);
  167. vecC = vldrhq_f16(inC);
  168. blkCnt--;
  169. }
  170. pBase += CMPLX_DIM * n1;
  171. }
  172. n1 = n2;
  173. n2 >>= 2u;
  174. iter = iter << 2;
  175. stage++;
  176. }
  177. /*
  178. * start of Last stage process
  179. */
  180. uint32x4_t vecScGathAddr = vld1q_u32((uint32_t*)strides);
  181. vecScGathAddr = vecScGathAddr + (uint32_t) pSrc;
  182. /* load scheduling */
  183. vecA = (f16x8_t)vldrwq_gather_base_wb_f32(&vecScGathAddr, 64);
  184. vecC = (f16x8_t)vldrwq_gather_base_f32(vecScGathAddr, 8);
  185. blkCnt = (fftLen >> 4);
  186. while (blkCnt > 0U)
  187. {
  188. vecSum0 = vecA + vecC; /* vecSum0 = vaddq(vecA, vecC) */
  189. vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */
  190. vecB = (f16x8_t)vldrwq_gather_base_f32(vecScGathAddr, 4);
  191. vecD = (f16x8_t)vldrwq_gather_base_f32(vecScGathAddr, 12);
  192. vecSum1 = vecB + vecD;
  193. vecDiff1 = vecB - vecD;
  194. /* pre-load for next iteration */
  195. vecA = (f16x8_t)vldrwq_gather_base_wb_f32(&vecScGathAddr, 64);
  196. vecC = (f16x8_t)vldrwq_gather_base_f32(vecScGathAddr, 8);
  197. vecTmp0 = vecSum0 + vecSum1;
  198. vstrwq_scatter_base_f32(vecScGathAddr, -64, (f32x4_t)vecTmp0);
  199. vecTmp0 = vecSum0 - vecSum1;
  200. vstrwq_scatter_base_f32(vecScGathAddr, -64 + 4, (f32x4_t)vecTmp0);
  201. vecTmp0 = MVE_CMPLX_SUB_A_ixB(vecDiff0, vecDiff1);
  202. vstrwq_scatter_base_f32(vecScGathAddr, -64 + 8, (f32x4_t)vecTmp0);
  203. vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1);
  204. vstrwq_scatter_base_f32(vecScGathAddr, -64 + 12, (f32x4_t)vecTmp0);
  205. blkCnt--;
  206. }
  207. /*
  208. * End of last stage process
  209. */
  210. }
  211. static void arm_cfft_radix4by2_f16_mve(const arm_cfft_instance_f16 * S, float16_t *pSrc, uint32_t fftLen)
  212. {
  213. float16_t const *pCoefVec;
  214. float16_t const *pCoef = S->pTwiddle;
  215. float16_t *pIn0, *pIn1;
  216. uint32_t n2;
  217. uint32_t blkCnt;
  218. f16x8_t vecIn0, vecIn1, vecSum, vecDiff;
  219. f16x8_t vecCmplxTmp, vecTw;
  220. n2 = fftLen >> 1;
  221. pIn0 = pSrc;
  222. pIn1 = pSrc + fftLen;
  223. pCoefVec = pCoef;
  224. blkCnt = n2 / 4;
  225. while (blkCnt > 0U)
  226. {
  227. vecIn0 = *(f16x8_t *) pIn0;
  228. vecIn1 = *(f16x8_t *) pIn1;
  229. vecTw = vld1q(pCoefVec);
  230. pCoefVec += 8;
  231. vecSum = vaddq(vecIn0, vecIn1);
  232. vecDiff = vsubq(vecIn0, vecIn1);
  233. vecCmplxTmp = MVE_CMPLX_MULT_FLT_Conj_AxB(vecTw, vecDiff);
  234. vst1q(pIn0, vecSum);
  235. pIn0 += 8;
  236. vst1q(pIn1, vecCmplxTmp);
  237. pIn1 += 8;
  238. blkCnt--;
  239. }
  240. _arm_radix4_butterfly_f16_mve(S, pSrc, n2);
  241. _arm_radix4_butterfly_f16_mve(S, pSrc + fftLen, n2);
  242. pIn0 = pSrc;
  243. }
  244. static void _arm_radix4_butterfly_inverse_f16_mve(const arm_cfft_instance_f16 * S,float16_t * pSrc, uint32_t fftLen, float16_t onebyfftLen)
  245. {
  246. f16x8_t vecTmp0, vecTmp1;
  247. f16x8_t vecSum0, vecDiff0, vecSum1, vecDiff1;
  248. f16x8_t vecA, vecB, vecC, vecD;
  249. uint32_t blkCnt;
  250. uint32_t n1, n2;
  251. uint32_t stage = 0;
  252. int32_t iter = 1;
  253. static const int32_t strides[4] = {
  254. ( 0 - 16) * (int32_t)sizeof(q31_t *),
  255. ( 4 - 16) * (int32_t)sizeof(q31_t *),
  256. ( 8 - 16) * (int32_t)sizeof(q31_t *),
  257. (12 - 16) * (int32_t)sizeof(q31_t *)
  258. };
  259. n2 = fftLen;
  260. n1 = n2;
  261. n2 >>= 2u;
  262. for (int k = fftLen / 4; k > 1; k >>= 2)
  263. {
  264. float16_t const *p_rearranged_twiddle_tab_stride1 =
  265. &S->rearranged_twiddle_stride1[
  266. S->rearranged_twiddle_tab_stride1_arr[stage]];
  267. float16_t const *p_rearranged_twiddle_tab_stride2 =
  268. &S->rearranged_twiddle_stride2[
  269. S->rearranged_twiddle_tab_stride2_arr[stage]];
  270. float16_t const *p_rearranged_twiddle_tab_stride3 =
  271. &S->rearranged_twiddle_stride3[
  272. S->rearranged_twiddle_tab_stride3_arr[stage]];
  273. float16_t * pBase = pSrc;
  274. for (int i = 0; i < iter; i++)
  275. {
  276. float16_t *inA = pBase;
  277. float16_t *inB = inA + n2 * CMPLX_DIM;
  278. float16_t *inC = inB + n2 * CMPLX_DIM;
  279. float16_t *inD = inC + n2 * CMPLX_DIM;
  280. float16_t const *pW1 = p_rearranged_twiddle_tab_stride1;
  281. float16_t const *pW2 = p_rearranged_twiddle_tab_stride2;
  282. float16_t const *pW3 = p_rearranged_twiddle_tab_stride3;
  283. f16x8_t vecW;
  284. blkCnt = n2 / 4;
  285. /*
  286. * load 2 f32 complex pair
  287. */
  288. vecA = vldrhq_f16(inA);
  289. vecC = vldrhq_f16(inC);
  290. while (blkCnt > 0U)
  291. {
  292. vecB = vldrhq_f16(inB);
  293. vecD = vldrhq_f16(inD);
  294. vecSum0 = vecA + vecC; /* vecSum0 = vaddq(vecA, vecC) */
  295. vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */
  296. vecSum1 = vecB + vecD;
  297. vecDiff1 = vecB - vecD;
  298. /*
  299. * [ 1 1 1 1 ] * [ A B C D ]' .* 1
  300. */
  301. vecTmp0 = vecSum0 + vecSum1;
  302. vst1q(inA, vecTmp0);
  303. inA += 8;
  304. /*
  305. * [ 1 -1 1 -1 ] * [ A B C D ]'
  306. */
  307. vecTmp0 = vecSum0 - vecSum1;
  308. /*
  309. * [ 1 -1 1 -1 ] * [ A B C D ]'.* W1
  310. */
  311. vecW = vld1q(pW2);
  312. pW2 += 8;
  313. vecTmp1 = MVE_CMPLX_MULT_FLT_AxB(vecW, vecTmp0);
  314. vst1q(inB, vecTmp1);
  315. inB += 8;
  316. /*
  317. * [ 1 -i -1 +i ] * [ A B C D ]'
  318. */
  319. vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1);
  320. /*
  321. * [ 1 -i -1 +i ] * [ A B C D ]'.* W2
  322. */
  323. vecW = vld1q(pW1);
  324. pW1 += 8;
  325. vecTmp1 = MVE_CMPLX_MULT_FLT_AxB(vecW, vecTmp0);
  326. vst1q(inC, vecTmp1);
  327. inC += 8;
  328. /*
  329. * [ 1 +i -1 -i ] * [ A B C D ]'
  330. */
  331. vecTmp0 = MVE_CMPLX_SUB_A_ixB(vecDiff0, vecDiff1);
  332. /*
  333. * [ 1 +i -1 -i ] * [ A B C D ]'.* W3
  334. */
  335. vecW = vld1q(pW3);
  336. pW3 += 8;
  337. vecTmp1 = MVE_CMPLX_MULT_FLT_AxB(vecW, vecTmp0);
  338. vst1q(inD, vecTmp1);
  339. inD += 8;
  340. vecA = vldrhq_f16(inA);
  341. vecC = vldrhq_f16(inC);
  342. blkCnt--;
  343. }
  344. pBase += CMPLX_DIM * n1;
  345. }
  346. n1 = n2;
  347. n2 >>= 2u;
  348. iter = iter << 2;
  349. stage++;
  350. }
  351. /*
  352. * start of Last stage process
  353. */
  354. uint32x4_t vecScGathAddr = vld1q_u32((uint32_t*)strides);
  355. vecScGathAddr = vecScGathAddr + (uint32_t) pSrc;
  356. /*
  357. * load scheduling
  358. */
  359. vecA = (f16x8_t)vldrwq_gather_base_wb_f32(&vecScGathAddr, 64);
  360. vecC = (f16x8_t)vldrwq_gather_base_f32(vecScGathAddr, 8);
  361. blkCnt = (fftLen >> 4);
  362. while (blkCnt > 0U)
  363. {
  364. vecSum0 = vecA + vecC; /* vecSum0 = vaddq(vecA, vecC) */
  365. vecDiff0 = vecA - vecC; /* vecSum0 = vsubq(vecA, vecC) */
  366. vecB = (f16x8_t)vldrwq_gather_base_f32(vecScGathAddr, 4);
  367. vecD = (f16x8_t)vldrwq_gather_base_f32(vecScGathAddr, 12);
  368. vecSum1 = vecB + vecD;
  369. vecDiff1 = vecB - vecD;
  370. vecA = (f16x8_t)vldrwq_gather_base_wb_f32(&vecScGathAddr, 64);
  371. vecC = (f16x8_t)vldrwq_gather_base_f32(vecScGathAddr, 8);
  372. vecTmp0 = vecSum0 + vecSum1;
  373. vecTmp0 = vecTmp0 * onebyfftLen;
  374. vstrwq_scatter_base_f32(vecScGathAddr, -64, (f32x4_t)vecTmp0);
  375. vecTmp0 = vecSum0 - vecSum1;
  376. vecTmp0 = vecTmp0 * onebyfftLen;
  377. vstrwq_scatter_base_f32(vecScGathAddr, -64 + 4, (f32x4_t)vecTmp0);
  378. vecTmp0 = MVE_CMPLX_ADD_A_ixB(vecDiff0, vecDiff1);
  379. vecTmp0 = vecTmp0 * onebyfftLen;
  380. vstrwq_scatter_base_f32(vecScGathAddr, -64 + 8, (f32x4_t)vecTmp0);
  381. vecTmp0 = MVE_CMPLX_SUB_A_ixB(vecDiff0, vecDiff1);
  382. vecTmp0 = vecTmp0 * onebyfftLen;
  383. vstrwq_scatter_base_f32(vecScGathAddr, -64 + 12, (f32x4_t)vecTmp0);
  384. blkCnt--;
  385. }
  386. /*
  387. * End of last stage process
  388. */
  389. }
  390. static void arm_cfft_radix4by2_inverse_f16_mve(const arm_cfft_instance_f16 * S,float16_t *pSrc, uint32_t fftLen)
  391. {
  392. float16_t const *pCoefVec;
  393. float16_t const *pCoef = S->pTwiddle;
  394. float16_t *pIn0, *pIn1;
  395. uint32_t n2;
  396. float16_t onebyfftLen = arm_inverse_fft_length_f16(fftLen);
  397. uint32_t blkCnt;
  398. f16x8_t vecIn0, vecIn1, vecSum, vecDiff;
  399. f16x8_t vecCmplxTmp, vecTw;
  400. n2 = fftLen >> 1;
  401. pIn0 = pSrc;
  402. pIn1 = pSrc + fftLen;
  403. pCoefVec = pCoef;
  404. blkCnt = n2 / 4;
  405. while (blkCnt > 0U)
  406. {
  407. vecIn0 = *(f16x8_t *) pIn0;
  408. vecIn1 = *(f16x8_t *) pIn1;
  409. vecTw = vld1q(pCoefVec);
  410. pCoefVec += 8;
  411. vecSum = vaddq(vecIn0, vecIn1);
  412. vecDiff = vsubq(vecIn0, vecIn1);
  413. vecCmplxTmp = MVE_CMPLX_MULT_FLT_AxB(vecTw, vecDiff);
  414. vst1q(pIn0, vecSum);
  415. pIn0 += 8;
  416. vst1q(pIn1, vecCmplxTmp);
  417. pIn1 += 8;
  418. blkCnt--;
  419. }
  420. _arm_radix4_butterfly_inverse_f16_mve(S, pSrc, n2, onebyfftLen);
  421. _arm_radix4_butterfly_inverse_f16_mve(S, pSrc + fftLen, n2, onebyfftLen);
  422. }
  423. /**
  424. @addtogroup ComplexFFTF16
  425. @{
  426. */
  427. /**
  428. @brief Processing function for the floating-point complex FFT.
  429. @param[in] S points to an instance of the floating-point CFFT structure
  430. @param[in,out] p1 points to the complex data buffer of size <code>2*fftLen</code>. Processing occurs in-place
  431. @param[in] ifftFlag flag that selects transform direction
  432. - value = 0: forward transform
  433. - value = 1: inverse transform
  434. @param[in] bitReverseFlag flag that enables / disables bit reversal of output
  435. - value = 0: disables bit reversal of output
  436. - value = 1: enables bit reversal of output
  437. */
  438. void arm_cfft_f16(
  439. const arm_cfft_instance_f16 * S,
  440. float16_t * pSrc,
  441. uint8_t ifftFlag,
  442. uint8_t bitReverseFlag)
  443. {
  444. uint32_t fftLen = S->fftLen;
  445. if (ifftFlag == 1U) {
  446. switch (fftLen) {
  447. case 16:
  448. case 64:
  449. case 256:
  450. case 1024:
  451. case 4096:
  452. _arm_radix4_butterfly_inverse_f16_mve(S, pSrc, fftLen, arm_inverse_fft_length_f16(S->fftLen));
  453. break;
  454. case 32:
  455. case 128:
  456. case 512:
  457. case 2048:
  458. arm_cfft_radix4by2_inverse_f16_mve(S, pSrc, fftLen);
  459. break;
  460. }
  461. } else {
  462. switch (fftLen) {
  463. case 16:
  464. case 64:
  465. case 256:
  466. case 1024:
  467. case 4096:
  468. _arm_radix4_butterfly_f16_mve(S, pSrc, fftLen);
  469. break;
  470. case 32:
  471. case 128:
  472. case 512:
  473. case 2048:
  474. arm_cfft_radix4by2_f16_mve(S, pSrc, fftLen);
  475. break;
  476. }
  477. }
  478. if (bitReverseFlag)
  479. {
  480. arm_bitreversal_16_inpl_mve((uint16_t*)pSrc, S->bitRevLength, S->pBitRevTable);
  481. }
  482. }
  483. #else
  484. #if defined(ARM_FLOAT16_SUPPORTED)
  485. extern void arm_bitreversal_16(
  486. uint16_t * pSrc,
  487. const uint16_t bitRevLen,
  488. const uint16_t * pBitRevTable);
  489. extern void arm_cfft_radix4by2_f16(
  490. float16_t * pSrc,
  491. uint32_t fftLen,
  492. const float16_t * pCoef);
  493. extern void arm_radix4_butterfly_f16(
  494. float16_t * pSrc,
  495. uint16_t fftLen,
  496. const float16_t * pCoef,
  497. uint16_t twidCoefModifier);
  498. /**
  499. @addtogroup ComplexFFTF16
  500. @{
  501. */
  502. /**
  503. @brief Processing function for the floating-point complex FFT.
  504. @param[in] S points to an instance of the floating-point CFFT structure
  505. @param[in,out] p1 points to the complex data buffer of size <code>2*fftLen</code>. Processing occurs in-place
  506. @param[in] ifftFlag flag that selects transform direction
  507. - value = 0: forward transform
  508. - value = 1: inverse transform
  509. @param[in] bitReverseFlag flag that enables / disables bit reversal of output
  510. - value = 0: disables bit reversal of output
  511. - value = 1: enables bit reversal of output
  512. */
  513. void arm_cfft_f16(
  514. const arm_cfft_instance_f16 * S,
  515. float16_t * p1,
  516. uint8_t ifftFlag,
  517. uint8_t bitReverseFlag)
  518. {
  519. uint32_t L = S->fftLen, l;
  520. float16_t invL, * pSrc;
  521. if (ifftFlag == 1U)
  522. {
  523. /* Conjugate input data */
  524. pSrc = p1 + 1;
  525. for(l=0; l<L; l++)
  526. {
  527. *pSrc = -(_Float16)*pSrc;
  528. pSrc += 2;
  529. }
  530. }
  531. switch (L)
  532. {
  533. case 16:
  534. case 64:
  535. case 256:
  536. case 1024:
  537. case 4096:
  538. arm_radix4_butterfly_f16 (p1, L, (float16_t*)S->pTwiddle, 1U);
  539. break;
  540. case 32:
  541. case 128:
  542. case 512:
  543. case 2048:
  544. arm_cfft_radix4by2_f16 ( p1, L, (float16_t*)S->pTwiddle);
  545. break;
  546. }
  547. if ( bitReverseFlag )
  548. arm_bitreversal_16((uint16_t*)p1, S->bitRevLength,(uint16_t*)S->pBitRevTable);
  549. if (ifftFlag == 1U)
  550. {
  551. invL = 1.0f16/(_Float16)L;
  552. /* Conjugate and scale output data */
  553. pSrc = p1;
  554. for(l=0; l<L; l++)
  555. {
  556. *pSrc++ *= (_Float16)invL ;
  557. *pSrc = -(_Float16)(*pSrc) * (_Float16)invL;
  558. pSrc++;
  559. }
  560. }
  561. }
  562. #endif /* if defined(ARM_FLOAT16_SUPPORTED) */
  563. #endif /* defined(ARM_MATH_MVEF) && !defined(ARM_MATH_AUTOVECTORIZE) */
  564. /**
  565. @} end of ComplexFFTF16 group
  566. */