rmt.c 55 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389
  1. // Copyright 2015-2020 Espressif Systems (Shanghai) PTE LTD
  2. //
  3. // Licensed under the Apache License, Version 2.0 (the "License");
  4. // you may not use this file except in compliance with the License.
  5. // You may obtain a copy of the License at
  6. //
  7. // http://www.apache.org/licenses/LICENSE-2.0
  8. //
  9. // Unless required by applicable law or agreed to in writing, software
  10. // distributed under the License is distributed on an "AS IS" BASIS,
  11. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. // See the License for the specific language governing permissions and
  13. // limitations under the License.
  14. #include <stdlib.h>
  15. #include <string.h>
  16. #include <sys/lock.h>
  17. #include <sys/cdefs.h>
  18. #include "esp_compiler.h"
  19. #include "esp_intr_alloc.h"
  20. #include "esp_log.h"
  21. #include "driver/gpio.h"
  22. #include "driver/periph_ctrl.h"
  23. #include "driver/rmt.h"
  24. #include "freertos/FreeRTOS.h"
  25. #include "freertos/task.h"
  26. #include "freertos/semphr.h"
  27. #include "freertos/ringbuf.h"
  28. #include "soc/soc_memory_layout.h"
  29. #include "soc/rmt_periph.h"
  30. #include "soc/rtc.h"
  31. #include "hal/rmt_hal.h"
  32. #include "hal/rmt_ll.h"
  33. #include "esp_rom_gpio.h"
  34. #define RMT_CHANNEL_ERROR_STR "RMT CHANNEL ERR"
  35. #define RMT_ADDR_ERROR_STR "RMT ADDRESS ERR"
  36. #define RMT_MEM_CNT_ERROR_STR "RMT MEM BLOCK NUM ERR"
  37. #define RMT_CARRIER_ERROR_STR "RMT CARRIER LEVEL ERR"
  38. #define RMT_MEM_OWNER_ERROR_STR "RMT MEM OWNER_ERR"
  39. #define RMT_BASECLK_ERROR_STR "RMT BASECLK ERR"
  40. #define RMT_WR_MEM_OVF_ERROR_STR "RMT WR MEM OVERFLOW"
  41. #define RMT_GPIO_ERROR_STR "RMT GPIO ERROR"
  42. #define RMT_MODE_ERROR_STR "RMT MODE ERROR"
  43. #define RMT_CLK_DIV_ERROR_STR "RMT CLK DIV ERR"
  44. #define RMT_DRIVER_ERROR_STR "RMT DRIVER ERR"
  45. #define RMT_DRIVER_LENGTH_ERROR_STR "RMT PARAM LEN ERROR"
  46. #define RMT_PSRAM_BUFFER_WARN_STR "Using buffer allocated from psram"
  47. #define RMT_TRANSLATOR_NULL_STR "RMT translator is null"
  48. #define RMT_TRANSLATOR_UNINIT_STR "RMT translator not init"
  49. #define RMT_PARAM_ERR_STR "RMT param error"
  50. static const char *RMT_TAG = "rmt";
  51. #define RMT_CHECK(a, str, ret_val, ...) \
  52. if (unlikely(!(a))) { \
  53. ESP_LOGE(RMT_TAG, "%s(%d): "str, __FUNCTION__, __LINE__, ##__VA_ARGS__); \
  54. return (ret_val); \
  55. }
  56. // Spinlock for protecting concurrent register-level access only
  57. #define RMT_ENTER_CRITICAL() portENTER_CRITICAL_SAFE(&(rmt_contex.rmt_spinlock))
  58. #define RMT_EXIT_CRITICAL() portEXIT_CRITICAL_SAFE(&(rmt_contex.rmt_spinlock))
  59. #define RMT_RX_CHANNEL_ENCODING_START (SOC_RMT_CHANNELS_PER_GROUP-SOC_RMT_TX_CANDIDATES_PER_GROUP)
  60. #define RMT_TX_CHANNEL_ENCODING_END (SOC_RMT_TX_CANDIDATES_PER_GROUP-1)
  61. #define RMT_IS_RX_CHANNEL(channel) ((channel) >= RMT_RX_CHANNEL_ENCODING_START)
  62. #define RMT_IS_TX_CHANNEL(channel) ((channel) <= RMT_TX_CHANNEL_ENCODING_END)
  63. #define RMT_DECODE_RX_CHANNEL(encode_chan) ((encode_chan - RMT_RX_CHANNEL_ENCODING_START))
  64. #define RMT_ENCODE_RX_CHANNEL(decode_chan) ((decode_chan + RMT_RX_CHANNEL_ENCODING_START))
  65. typedef struct {
  66. rmt_hal_context_t hal;
  67. _lock_t rmt_driver_isr_lock;
  68. portMUX_TYPE rmt_spinlock; // Mutex lock for protecting concurrent register/unregister of RMT channels' ISR
  69. rmt_isr_handle_t rmt_driver_intr_handle;
  70. rmt_tx_end_callback_t rmt_tx_end_callback;// Event called when transmission is ended
  71. uint8_t rmt_driver_channels; // Bitmask of installed drivers' channels
  72. bool rmt_module_enabled;
  73. uint32_t synchro_channel_mask; // Bitmap of channels already added in the synchronous group
  74. } rmt_contex_t;
  75. typedef struct {
  76. size_t tx_offset;
  77. size_t tx_len_rem;
  78. size_t tx_sub_len;
  79. bool translator;
  80. bool wait_done; //Mark whether wait tx done.
  81. rmt_channel_t channel;
  82. const rmt_item32_t *tx_data;
  83. xSemaphoreHandle tx_sem;
  84. #if CONFIG_SPIRAM_USE_MALLOC
  85. int intr_alloc_flags;
  86. StaticSemaphore_t tx_sem_buffer;
  87. #endif
  88. rmt_item32_t *tx_buf;
  89. RingbufHandle_t rx_buf;
  90. #if SOC_RMT_SUPPORT_RX_PINGPONG
  91. rmt_item32_t *rx_item_buf;
  92. uint32_t rx_item_buf_size;
  93. uint32_t rx_item_len;
  94. int rx_item_start_idx;
  95. #endif
  96. sample_to_rmt_t sample_to_rmt;
  97. void *tx_context;
  98. size_t sample_size_remain;
  99. const uint8_t *sample_cur;
  100. } rmt_obj_t;
  101. static rmt_contex_t rmt_contex = {
  102. .hal.regs = RMT_LL_HW_BASE,
  103. .hal.mem = RMT_LL_MEM_BASE,
  104. .rmt_spinlock = portMUX_INITIALIZER_UNLOCKED,
  105. .rmt_driver_intr_handle = NULL,
  106. .rmt_tx_end_callback = {
  107. .function = NULL,
  108. },
  109. .rmt_driver_channels = 0,
  110. .rmt_module_enabled = false,
  111. .synchro_channel_mask = 0
  112. };
  113. static rmt_obj_t *p_rmt_obj[RMT_CHANNEL_MAX] = {0};
  114. #if SOC_RMT_SOURCE_CLK_INDEPENDENT
  115. static uint32_t s_rmt_source_clock_hz[RMT_CHANNEL_MAX];
  116. #else
  117. static uint32_t s_rmt_source_clock_hz;
  118. #endif
  119. //Enable RMT module
  120. static void rmt_module_enable(void)
  121. {
  122. RMT_ENTER_CRITICAL();
  123. if (rmt_contex.rmt_module_enabled == false) {
  124. periph_module_reset(rmt_periph_signals.module);
  125. periph_module_enable(rmt_periph_signals.module);
  126. rmt_contex.rmt_module_enabled = true;
  127. }
  128. RMT_EXIT_CRITICAL();
  129. }
  130. //Disable RMT module
  131. static void rmt_module_disable(void)
  132. {
  133. RMT_ENTER_CRITICAL();
  134. if (rmt_contex.rmt_module_enabled == true) {
  135. periph_module_disable(rmt_periph_signals.module);
  136. rmt_contex.rmt_module_enabled = false;
  137. }
  138. RMT_EXIT_CRITICAL();
  139. }
  140. esp_err_t rmt_set_clk_div(rmt_channel_t channel, uint8_t div_cnt)
  141. {
  142. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  143. RMT_ENTER_CRITICAL();
  144. if (RMT_IS_RX_CHANNEL(channel)) {
  145. rmt_ll_rx_set_channel_clock_div(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), div_cnt);
  146. } else {
  147. rmt_ll_tx_set_channel_clock_div(rmt_contex.hal.regs, channel, div_cnt);
  148. }
  149. RMT_EXIT_CRITICAL();
  150. return ESP_OK;
  151. }
  152. esp_err_t rmt_get_clk_div(rmt_channel_t channel, uint8_t *div_cnt)
  153. {
  154. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  155. RMT_CHECK(div_cnt != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  156. RMT_ENTER_CRITICAL();
  157. if (RMT_IS_RX_CHANNEL(channel)) {
  158. *div_cnt = (uint8_t)rmt_ll_rx_get_channel_clock_div(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  159. } else {
  160. *div_cnt = (uint8_t)rmt_ll_tx_get_channel_clock_div(rmt_contex.hal.regs, channel);
  161. }
  162. RMT_EXIT_CRITICAL();
  163. return ESP_OK;
  164. }
  165. esp_err_t rmt_set_rx_idle_thresh(rmt_channel_t channel, uint16_t thresh)
  166. {
  167. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  168. RMT_ENTER_CRITICAL();
  169. rmt_ll_rx_set_idle_thres(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), thresh);
  170. RMT_EXIT_CRITICAL();
  171. return ESP_OK;
  172. }
  173. esp_err_t rmt_get_rx_idle_thresh(rmt_channel_t channel, uint16_t *thresh)
  174. {
  175. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  176. RMT_CHECK(thresh != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  177. RMT_ENTER_CRITICAL();
  178. *thresh = (uint16_t)rmt_ll_rx_get_idle_thres(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  179. RMT_EXIT_CRITICAL();
  180. return ESP_OK;
  181. }
  182. esp_err_t rmt_set_mem_block_num(rmt_channel_t channel, uint8_t rmt_mem_num)
  183. {
  184. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  185. RMT_CHECK(rmt_mem_num <= RMT_CHANNEL_MAX - channel, RMT_MEM_CNT_ERROR_STR, ESP_ERR_INVALID_ARG);
  186. RMT_ENTER_CRITICAL();
  187. if (RMT_IS_RX_CHANNEL(channel)) {
  188. rmt_ll_rx_set_mem_blocks(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), rmt_mem_num);
  189. } else {
  190. rmt_ll_tx_set_mem_blocks(rmt_contex.hal.regs, channel, rmt_mem_num);
  191. }
  192. RMT_EXIT_CRITICAL();
  193. return ESP_OK;
  194. }
  195. esp_err_t rmt_get_mem_block_num(rmt_channel_t channel, uint8_t *rmt_mem_num)
  196. {
  197. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  198. RMT_CHECK(rmt_mem_num != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  199. RMT_ENTER_CRITICAL();
  200. if (RMT_IS_RX_CHANNEL(channel)) {
  201. *rmt_mem_num = (uint8_t)rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  202. } else {
  203. *rmt_mem_num = (uint8_t)rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel);
  204. }
  205. RMT_EXIT_CRITICAL();
  206. return ESP_OK;
  207. }
  208. esp_err_t rmt_set_tx_carrier(rmt_channel_t channel, bool carrier_en, uint16_t high_level, uint16_t low_level,
  209. rmt_carrier_level_t carrier_level)
  210. {
  211. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  212. RMT_CHECK(carrier_level < RMT_CARRIER_LEVEL_MAX, RMT_CARRIER_ERROR_STR, ESP_ERR_INVALID_ARG);
  213. RMT_ENTER_CRITICAL();
  214. rmt_ll_tx_set_carrier_high_low_ticks(rmt_contex.hal.regs, channel, high_level, low_level);
  215. rmt_ll_tx_set_carrier_level(rmt_contex.hal.regs, channel, carrier_level);
  216. rmt_ll_tx_enable_carrier_modulation(rmt_contex.hal.regs, channel, carrier_en);
  217. RMT_EXIT_CRITICAL();
  218. return ESP_OK;
  219. }
  220. esp_err_t rmt_set_mem_pd(rmt_channel_t channel, bool pd_en)
  221. {
  222. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  223. RMT_ENTER_CRITICAL();
  224. rmt_ll_power_down_mem(rmt_contex.hal.regs, pd_en);
  225. RMT_EXIT_CRITICAL();
  226. return ESP_OK;
  227. }
  228. esp_err_t rmt_get_mem_pd(rmt_channel_t channel, bool *pd_en)
  229. {
  230. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  231. RMT_ENTER_CRITICAL();
  232. *pd_en = rmt_ll_is_mem_power_down(rmt_contex.hal.regs);
  233. RMT_EXIT_CRITICAL();
  234. return ESP_OK;
  235. }
  236. esp_err_t rmt_tx_start(rmt_channel_t channel, bool tx_idx_rst)
  237. {
  238. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  239. RMT_ENTER_CRITICAL();
  240. if (tx_idx_rst) {
  241. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  242. }
  243. rmt_ll_clear_tx_end_interrupt(rmt_contex.hal.regs, channel);
  244. // enable tx end interrupt in non-loop mode
  245. if (!rmt_ll_is_tx_loop_enabled(rmt_contex.hal.regs, channel)) {
  246. rmt_ll_enable_tx_end_interrupt(rmt_contex.hal.regs, channel, true);
  247. } else {
  248. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  249. rmt_ll_tx_reset_loop(rmt_contex.hal.regs, channel);
  250. rmt_ll_tx_enable_loop_count(rmt_contex.hal.regs, channel, true);
  251. rmt_ll_clear_tx_loop_interrupt(rmt_contex.hal.regs, channel);
  252. rmt_ll_enable_tx_loop_interrupt(rmt_contex.hal.regs, channel, true);
  253. #endif
  254. }
  255. rmt_ll_tx_start(rmt_contex.hal.regs, channel);
  256. RMT_EXIT_CRITICAL();
  257. return ESP_OK;
  258. }
  259. esp_err_t rmt_tx_stop(rmt_channel_t channel)
  260. {
  261. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  262. RMT_ENTER_CRITICAL();
  263. rmt_ll_tx_stop(rmt_contex.hal.regs, channel);
  264. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  265. RMT_EXIT_CRITICAL();
  266. return ESP_OK;
  267. }
  268. esp_err_t rmt_rx_start(rmt_channel_t channel, bool rx_idx_rst)
  269. {
  270. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  271. RMT_ENTER_CRITICAL();
  272. rmt_ll_rx_enable(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  273. if (rx_idx_rst) {
  274. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  275. }
  276. rmt_ll_clear_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  277. rmt_ll_enable_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), true);
  278. #if SOC_RMT_SUPPORT_RX_PINGPONG
  279. const uint32_t item_block_len = rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel)) * RMT_MEM_ITEM_NUM;
  280. p_rmt_obj[channel]->rx_item_start_idx = 0;
  281. p_rmt_obj[channel]->rx_item_len = 0;
  282. rmt_set_rx_thr_intr_en(channel, true, item_block_len / 2);
  283. #endif
  284. rmt_ll_rx_enable(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), true);
  285. RMT_EXIT_CRITICAL();
  286. return ESP_OK;
  287. }
  288. esp_err_t rmt_rx_stop(rmt_channel_t channel)
  289. {
  290. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  291. RMT_ENTER_CRITICAL();
  292. rmt_ll_enable_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  293. rmt_ll_rx_enable(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  294. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  295. #if SOC_RMT_SUPPORT_RX_PINGPONG
  296. rmt_ll_enable_rx_thres_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  297. #endif
  298. RMT_EXIT_CRITICAL();
  299. return ESP_OK;
  300. }
  301. esp_err_t rmt_tx_memory_reset(rmt_channel_t channel)
  302. {
  303. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  304. RMT_ENTER_CRITICAL();
  305. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  306. RMT_EXIT_CRITICAL();
  307. return ESP_OK;
  308. }
  309. esp_err_t rmt_rx_memory_reset(rmt_channel_t channel)
  310. {
  311. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  312. RMT_ENTER_CRITICAL();
  313. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  314. RMT_EXIT_CRITICAL();
  315. return ESP_OK;
  316. }
  317. esp_err_t rmt_set_memory_owner(rmt_channel_t channel, rmt_mem_owner_t owner)
  318. {
  319. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  320. RMT_CHECK(owner < RMT_MEM_OWNER_MAX, RMT_MEM_OWNER_ERROR_STR, ESP_ERR_INVALID_ARG);
  321. RMT_ENTER_CRITICAL();
  322. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), owner);
  323. RMT_EXIT_CRITICAL();
  324. return ESP_OK;
  325. }
  326. esp_err_t rmt_get_memory_owner(rmt_channel_t channel, rmt_mem_owner_t *owner)
  327. {
  328. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  329. RMT_CHECK(owner != NULL, RMT_MEM_OWNER_ERROR_STR, ESP_ERR_INVALID_ARG);
  330. RMT_ENTER_CRITICAL();
  331. *owner = (rmt_mem_owner_t)rmt_ll_rx_get_mem_owner(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  332. RMT_EXIT_CRITICAL();
  333. return ESP_OK;
  334. }
  335. esp_err_t rmt_set_tx_loop_mode(rmt_channel_t channel, bool loop_en)
  336. {
  337. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  338. RMT_ENTER_CRITICAL();
  339. rmt_ll_tx_enable_loop(rmt_contex.hal.regs, channel, loop_en);
  340. RMT_EXIT_CRITICAL();
  341. return ESP_OK;
  342. }
  343. esp_err_t rmt_get_tx_loop_mode(rmt_channel_t channel, bool *loop_en)
  344. {
  345. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  346. RMT_ENTER_CRITICAL();
  347. *loop_en = rmt_ll_is_tx_loop_enabled(rmt_contex.hal.regs, channel);
  348. RMT_EXIT_CRITICAL();
  349. return ESP_OK;
  350. }
  351. esp_err_t rmt_set_rx_filter(rmt_channel_t channel, bool rx_filter_en, uint8_t thresh)
  352. {
  353. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  354. RMT_ENTER_CRITICAL();
  355. rmt_ll_rx_enable_filter(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), rx_filter_en);
  356. rmt_ll_rx_set_filter_thres(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), thresh);
  357. RMT_EXIT_CRITICAL();
  358. return ESP_OK;
  359. }
  360. esp_err_t rmt_set_source_clk(rmt_channel_t channel, rmt_source_clk_t base_clk)
  361. {
  362. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  363. RMT_CHECK(base_clk < RMT_BASECLK_MAX, RMT_BASECLK_ERROR_STR, ESP_ERR_INVALID_ARG);
  364. RMT_ENTER_CRITICAL();
  365. rmt_ll_set_group_clock_src(rmt_contex.hal.regs, channel, base_clk, 0, 0, 0);
  366. RMT_EXIT_CRITICAL();
  367. return ESP_OK;
  368. }
  369. esp_err_t rmt_get_source_clk(rmt_channel_t channel, rmt_source_clk_t *src_clk)
  370. {
  371. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  372. RMT_ENTER_CRITICAL();
  373. *src_clk = (rmt_source_clk_t)rmt_ll_get_group_clock_src(rmt_contex.hal.regs, channel);
  374. RMT_EXIT_CRITICAL();
  375. return ESP_OK;
  376. }
  377. esp_err_t rmt_set_idle_level(rmt_channel_t channel, bool idle_out_en, rmt_idle_level_t level)
  378. {
  379. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  380. RMT_CHECK(level < RMT_IDLE_LEVEL_MAX, "RMT IDLE LEVEL ERR", ESP_ERR_INVALID_ARG);
  381. RMT_ENTER_CRITICAL();
  382. rmt_ll_tx_enable_idle(rmt_contex.hal.regs, channel, idle_out_en);
  383. rmt_ll_tx_set_idle_level(rmt_contex.hal.regs, channel, level);
  384. RMT_EXIT_CRITICAL();
  385. return ESP_OK;
  386. }
  387. esp_err_t rmt_get_idle_level(rmt_channel_t channel, bool *idle_out_en, rmt_idle_level_t *level)
  388. {
  389. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  390. RMT_ENTER_CRITICAL();
  391. *idle_out_en = rmt_ll_is_tx_idle_enabled(rmt_contex.hal.regs, channel);
  392. *level = rmt_ll_tx_get_idle_level(rmt_contex.hal.regs, channel);
  393. RMT_EXIT_CRITICAL();
  394. return ESP_OK;
  395. }
  396. esp_err_t rmt_get_status(rmt_channel_t channel, uint32_t *status)
  397. {
  398. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  399. RMT_ENTER_CRITICAL();
  400. if (RMT_IS_RX_CHANNEL(channel)) {
  401. *status = rmt_ll_rx_get_channel_status(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  402. } else {
  403. *status = rmt_ll_tx_get_channel_status(rmt_contex.hal.regs, channel);
  404. }
  405. RMT_EXIT_CRITICAL();
  406. return ESP_OK;
  407. }
  408. void rmt_set_intr_enable_mask(uint32_t mask)
  409. {
  410. RMT_ENTER_CRITICAL();
  411. rmt_ll_set_intr_enable_mask(mask);
  412. RMT_EXIT_CRITICAL();
  413. }
  414. void rmt_clr_intr_enable_mask(uint32_t mask)
  415. {
  416. RMT_ENTER_CRITICAL();
  417. rmt_ll_clr_intr_enable_mask(mask);
  418. RMT_EXIT_CRITICAL();
  419. }
  420. esp_err_t rmt_set_rx_intr_en(rmt_channel_t channel, bool en)
  421. {
  422. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  423. RMT_ENTER_CRITICAL();
  424. rmt_ll_enable_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), en);
  425. RMT_EXIT_CRITICAL();
  426. return ESP_OK;
  427. }
  428. #if SOC_RMT_SUPPORT_RX_PINGPONG
  429. esp_err_t rmt_set_rx_thr_intr_en(rmt_channel_t channel, bool en, uint16_t evt_thresh)
  430. {
  431. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  432. if (en) {
  433. uint32_t item_block_len = rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel)) * RMT_MEM_ITEM_NUM;
  434. RMT_CHECK(evt_thresh <= item_block_len, "RMT EVT THRESH ERR", ESP_ERR_INVALID_ARG);
  435. RMT_ENTER_CRITICAL();
  436. rmt_ll_rx_set_limit(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), evt_thresh);
  437. rmt_ll_enable_rx_thres_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), true);
  438. RMT_EXIT_CRITICAL();
  439. } else {
  440. RMT_ENTER_CRITICAL();
  441. rmt_ll_enable_rx_thres_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  442. RMT_EXIT_CRITICAL();
  443. }
  444. return ESP_OK;
  445. }
  446. #endif
  447. esp_err_t rmt_set_err_intr_en(rmt_channel_t channel, bool en)
  448. {
  449. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  450. RMT_ENTER_CRITICAL();
  451. if (RMT_IS_RX_CHANNEL(channel)) {
  452. rmt_ll_enable_rx_err_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), en);
  453. } else {
  454. rmt_ll_enable_tx_err_interrupt(rmt_contex.hal.regs, channel, en);
  455. }
  456. RMT_EXIT_CRITICAL();
  457. return ESP_OK;
  458. }
  459. esp_err_t rmt_set_tx_intr_en(rmt_channel_t channel, bool en)
  460. {
  461. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  462. RMT_ENTER_CRITICAL();
  463. rmt_ll_enable_tx_end_interrupt(rmt_contex.hal.regs, channel, en);
  464. RMT_EXIT_CRITICAL();
  465. return ESP_OK;
  466. }
  467. esp_err_t rmt_set_tx_thr_intr_en(rmt_channel_t channel, bool en, uint16_t evt_thresh)
  468. {
  469. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  470. if (en) {
  471. uint32_t item_block_len = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel) * RMT_MEM_ITEM_NUM;
  472. RMT_CHECK(evt_thresh <= item_block_len, "RMT EVT THRESH ERR", ESP_ERR_INVALID_ARG);
  473. RMT_ENTER_CRITICAL();
  474. rmt_ll_tx_set_limit(rmt_contex.hal.regs, channel, evt_thresh);
  475. rmt_ll_enable_tx_thres_interrupt(rmt_contex.hal.regs, channel, true);
  476. RMT_EXIT_CRITICAL();
  477. } else {
  478. RMT_ENTER_CRITICAL();
  479. rmt_ll_enable_tx_thres_interrupt(rmt_contex.hal.regs, channel, false);
  480. RMT_EXIT_CRITICAL();
  481. }
  482. return ESP_OK;
  483. }
  484. esp_err_t rmt_set_gpio(rmt_channel_t channel, rmt_mode_t mode, gpio_num_t gpio_num, bool invert_signal)
  485. {
  486. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  487. RMT_CHECK(mode < RMT_MODE_MAX, RMT_MODE_ERROR_STR, ESP_ERR_INVALID_ARG);
  488. RMT_CHECK(((GPIO_IS_VALID_GPIO(gpio_num) && (mode == RMT_MODE_RX)) ||
  489. (GPIO_IS_VALID_OUTPUT_GPIO(gpio_num) && (mode == RMT_MODE_TX))),
  490. RMT_GPIO_ERROR_STR, ESP_ERR_INVALID_ARG);
  491. PIN_FUNC_SELECT(GPIO_PIN_MUX_REG[gpio_num], PIN_FUNC_GPIO);
  492. if (mode == RMT_MODE_TX) {
  493. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  494. gpio_set_direction(gpio_num, GPIO_MODE_OUTPUT);
  495. esp_rom_gpio_connect_out_signal(gpio_num, rmt_periph_signals.channels[channel].tx_sig, invert_signal, 0);
  496. } else {
  497. RMT_CHECK(RMT_IS_RX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  498. gpio_set_direction(gpio_num, GPIO_MODE_INPUT);
  499. esp_rom_gpio_connect_in_signal(gpio_num, rmt_periph_signals.channels[channel].rx_sig, invert_signal);
  500. }
  501. return ESP_OK;
  502. }
  503. esp_err_t rmt_set_pin(rmt_channel_t channel, rmt_mode_t mode, gpio_num_t gpio_num)
  504. {
  505. // only for backword compatibility
  506. return rmt_set_gpio(channel, mode, gpio_num, false);
  507. }
  508. static bool rmt_is_channel_number_valid(rmt_channel_t channel, uint8_t mode)
  509. {
  510. // RX mode
  511. if (mode == RMT_MODE_RX) {
  512. return RMT_IS_RX_CHANNEL(channel) && (channel < RMT_CHANNEL_MAX);
  513. }
  514. // TX mode
  515. return (channel >= 0) && RMT_IS_TX_CHANNEL(channel);
  516. }
  517. static esp_err_t rmt_internal_config(rmt_dev_t *dev, const rmt_config_t *rmt_param)
  518. {
  519. uint8_t mode = rmt_param->rmt_mode;
  520. uint8_t channel = rmt_param->channel;
  521. uint8_t gpio_num = rmt_param->gpio_num;
  522. uint8_t mem_cnt = rmt_param->mem_block_num;
  523. uint8_t clk_div = rmt_param->clk_div;
  524. uint32_t carrier_freq_hz = rmt_param->tx_config.carrier_freq_hz;
  525. bool carrier_en = rmt_param->tx_config.carrier_en;
  526. uint32_t rmt_source_clk_hz;
  527. RMT_CHECK(rmt_is_channel_number_valid(channel, mode), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  528. RMT_CHECK((mem_cnt + channel <= 8 && mem_cnt > 0), RMT_MEM_CNT_ERROR_STR, ESP_ERR_INVALID_ARG);
  529. RMT_CHECK((clk_div > 0), RMT_CLK_DIV_ERROR_STR, ESP_ERR_INVALID_ARG);
  530. if (mode == RMT_MODE_TX) {
  531. RMT_CHECK((!carrier_en || carrier_freq_hz > 0), "RMT carrier frequency can't be zero", ESP_ERR_INVALID_ARG);
  532. }
  533. RMT_ENTER_CRITICAL();
  534. rmt_ll_enable_mem_access(dev, true);
  535. if (rmt_param->flags & RMT_CHANNEL_FLAGS_AWARE_DFS) {
  536. #if SOC_RMT_SUPPORT_XTAL
  537. // clock src: XTAL_CLK
  538. rmt_source_clk_hz = rtc_clk_xtal_freq_get() * 1000000;
  539. rmt_ll_set_group_clock_src(dev, channel, RMT_BASECLK_XTAL, 0, 0, 0);
  540. #elif SOC_RMT_SUPPORT_REF_TICK
  541. // clock src: REF_CLK
  542. rmt_source_clk_hz = REF_CLK_FREQ;
  543. rmt_ll_set_group_clock_src(dev, channel, RMT_BASECLK_REF, 0, 0, 0);
  544. #endif
  545. } else {
  546. // clock src: APB_CLK
  547. rmt_source_clk_hz = APB_CLK_FREQ;
  548. rmt_ll_set_group_clock_src(dev, channel, RMT_BASECLK_APB, 0, 0, 0);
  549. }
  550. RMT_EXIT_CRITICAL();
  551. #if SOC_RMT_SOURCE_CLK_INDEPENDENT
  552. s_rmt_source_clock_hz[channel] = rmt_source_clk_hz;
  553. #else
  554. if (s_rmt_source_clock_hz && rmt_source_clk_hz != s_rmt_source_clock_hz) {
  555. ESP_LOGW(RMT_TAG, "RMT clock source has been configured to %d by other channel, now reconfigure it to %d", s_rmt_source_clock_hz, rmt_source_clk_hz);
  556. }
  557. s_rmt_source_clock_hz = rmt_source_clk_hz;
  558. #endif
  559. ESP_LOGD(RMT_TAG, "rmt_source_clk_hz: %d\n", rmt_source_clk_hz);
  560. if (mode == RMT_MODE_TX) {
  561. uint16_t carrier_duty_percent = rmt_param->tx_config.carrier_duty_percent;
  562. uint8_t carrier_level = rmt_param->tx_config.carrier_level;
  563. uint8_t idle_level = rmt_param->tx_config.idle_level;
  564. RMT_ENTER_CRITICAL();
  565. rmt_ll_tx_set_channel_clock_div(dev, channel, clk_div);
  566. rmt_ll_tx_set_mem_blocks(dev, channel, mem_cnt);
  567. rmt_ll_tx_reset_pointer(dev, channel);
  568. rmt_ll_tx_enable_loop(dev, channel, rmt_param->tx_config.loop_en);
  569. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  570. if (rmt_param->tx_config.loop_en) {
  571. rmt_ll_tx_set_loop_count(dev, channel, rmt_param->tx_config.loop_count);
  572. }
  573. #endif
  574. /* always enable tx ping-pong */
  575. rmt_ll_tx_enable_pingpong(dev, channel, true);
  576. /*Set idle level */
  577. rmt_ll_tx_enable_idle(dev, channel, rmt_param->tx_config.idle_output_en);
  578. rmt_ll_tx_set_idle_level(dev, channel, idle_level);
  579. /*Set carrier*/
  580. rmt_ll_tx_enable_carrier_modulation(dev, channel, carrier_en);
  581. if (carrier_en) {
  582. uint32_t duty_div, duty_h, duty_l;
  583. duty_div = rmt_source_clk_hz / carrier_freq_hz;
  584. duty_h = duty_div * carrier_duty_percent / 100;
  585. duty_l = duty_div - duty_h;
  586. rmt_ll_tx_set_carrier_level(dev, channel, carrier_level);
  587. rmt_ll_tx_set_carrier_high_low_ticks(dev, channel, duty_h, duty_l);
  588. } else {
  589. rmt_ll_tx_set_carrier_level(dev, channel, 0);
  590. rmt_ll_tx_set_carrier_high_low_ticks(dev, channel, 0, 0);
  591. }
  592. RMT_EXIT_CRITICAL();
  593. ESP_LOGD(RMT_TAG, "Rmt Tx Channel %u|Gpio %u|Sclk_Hz %u|Div %u|Carrier_Hz %u|Duty %u",
  594. channel, gpio_num, rmt_source_clk_hz, clk_div, carrier_freq_hz, carrier_duty_percent);
  595. } else if (RMT_MODE_RX == mode) {
  596. uint8_t filter_cnt = rmt_param->rx_config.filter_ticks_thresh;
  597. uint16_t threshold = rmt_param->rx_config.idle_threshold;
  598. RMT_ENTER_CRITICAL();
  599. rmt_ll_rx_set_channel_clock_div(dev, RMT_DECODE_RX_CHANNEL(channel), clk_div);
  600. rmt_ll_rx_set_mem_blocks(dev, RMT_DECODE_RX_CHANNEL(channel), mem_cnt);
  601. rmt_ll_rx_reset_pointer(dev, RMT_DECODE_RX_CHANNEL(channel));
  602. rmt_ll_rx_set_mem_owner(dev, RMT_DECODE_RX_CHANNEL(channel), RMT_MEM_OWNER_HW);
  603. /*Set idle threshold*/
  604. rmt_ll_rx_set_idle_thres(dev, RMT_DECODE_RX_CHANNEL(channel), threshold);
  605. /* Set RX filter */
  606. rmt_ll_rx_set_filter_thres(dev, RMT_DECODE_RX_CHANNEL(channel), filter_cnt);
  607. rmt_ll_rx_enable_filter(dev, RMT_DECODE_RX_CHANNEL(channel), rmt_param->rx_config.filter_en);
  608. #if SOC_RMT_SUPPORT_RX_PINGPONG
  609. /* always enable rx ping-pong */
  610. rmt_ll_rx_enable_pingpong(dev, RMT_DECODE_RX_CHANNEL(channel), true);
  611. #endif
  612. #if SOC_RMT_SUPPORT_RX_DEMODULATION
  613. rmt_ll_rx_enable_carrier_demodulation(dev, RMT_DECODE_RX_CHANNEL(channel), rmt_param->rx_config.rm_carrier);
  614. if (rmt_param->rx_config.rm_carrier) {
  615. uint32_t duty_total = rmt_source_clk_hz / rmt_ll_rx_get_channel_clock_div(dev, RMT_DECODE_RX_CHANNEL(channel)) / rmt_param->rx_config.carrier_freq_hz;
  616. uint32_t duty_high = duty_total * rmt_param->rx_config.carrier_duty_percent / 100;
  617. // there could be residual in timing the carrier pulse, so double enlarge the theoretical value
  618. rmt_ll_rx_set_carrier_high_low_ticks(dev, RMT_DECODE_RX_CHANNEL(channel), duty_high * 2, (duty_total - duty_high) * 2);
  619. rmt_ll_rx_set_carrier_level(dev, RMT_DECODE_RX_CHANNEL(channel), rmt_param->rx_config.carrier_level);
  620. }
  621. #endif
  622. RMT_EXIT_CRITICAL();
  623. ESP_LOGD(RMT_TAG, "Rmt Rx Channel %u|Gpio %u|Sclk_Hz %u|Div %u|Thresold %u|Filter %u",
  624. channel, gpio_num, rmt_source_clk_hz, clk_div, threshold, filter_cnt);
  625. }
  626. return ESP_OK;
  627. }
  628. esp_err_t rmt_config(const rmt_config_t *rmt_param)
  629. {
  630. rmt_module_enable();
  631. RMT_CHECK(rmt_set_gpio(rmt_param->channel, rmt_param->rmt_mode, rmt_param->gpio_num, rmt_param->flags & RMT_CHANNEL_FLAGS_INVERT_SIG) == ESP_OK,
  632. "set gpio for RMT driver failed", ESP_ERR_INVALID_ARG);
  633. RMT_CHECK(rmt_internal_config(&RMT, rmt_param) == ESP_OK,
  634. "initialize RMT driver failed", ESP_ERR_INVALID_ARG);
  635. return ESP_OK;
  636. }
  637. static void IRAM_ATTR rmt_fill_memory(rmt_channel_t channel, const rmt_item32_t *item,
  638. uint16_t item_num, uint16_t mem_offset)
  639. {
  640. RMT_ENTER_CRITICAL();
  641. rmt_ll_write_memory(rmt_contex.hal.mem, channel, item, item_num, mem_offset);
  642. RMT_EXIT_CRITICAL();
  643. }
  644. esp_err_t rmt_fill_tx_items(rmt_channel_t channel, const rmt_item32_t *item, uint16_t item_num, uint16_t mem_offset)
  645. {
  646. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, (0));
  647. RMT_CHECK((item != NULL), RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  648. RMT_CHECK((item_num > 0), RMT_DRIVER_LENGTH_ERROR_STR, ESP_ERR_INVALID_ARG);
  649. /*Each block has 64 x 32 bits of data*/
  650. uint8_t mem_cnt = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel);
  651. RMT_CHECK((mem_cnt * RMT_MEM_ITEM_NUM >= item_num), RMT_WR_MEM_OVF_ERROR_STR, ESP_ERR_INVALID_ARG);
  652. rmt_fill_memory(channel, item, item_num, mem_offset);
  653. return ESP_OK;
  654. }
  655. esp_err_t rmt_isr_register(void (*fn)(void *), void *arg, int intr_alloc_flags, rmt_isr_handle_t *handle)
  656. {
  657. RMT_CHECK((fn != NULL), RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  658. RMT_CHECK(rmt_contex.rmt_driver_channels == 0, "RMT driver installed, can not install generic ISR handler", ESP_FAIL);
  659. return esp_intr_alloc(rmt_periph_signals.irq, intr_alloc_flags, fn, arg, handle);
  660. }
  661. esp_err_t rmt_isr_deregister(rmt_isr_handle_t handle)
  662. {
  663. return esp_intr_free(handle);
  664. }
  665. static int IRAM_ATTR rmt_rx_get_mem_len_in_isr(rmt_channel_t channel)
  666. {
  667. int block_num = rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, channel);
  668. int item_block_len = block_num * RMT_MEM_ITEM_NUM;
  669. volatile rmt_item32_t *data = (rmt_item32_t *)RMTMEM.chan[RMT_ENCODE_RX_CHANNEL(channel)].data32;
  670. int idx;
  671. for (idx = 0; idx < item_block_len; idx++) {
  672. if (data[idx].duration0 == 0) {
  673. return idx;
  674. } else if (data[idx].duration1 == 0) {
  675. return idx + 1;
  676. }
  677. }
  678. return idx;
  679. }
  680. static void IRAM_ATTR rmt_driver_isr_default(void *arg)
  681. {
  682. uint32_t status = 0;
  683. rmt_item32_t volatile *addr = NULL;
  684. uint8_t channel = 0;
  685. rmt_hal_context_t *hal = (rmt_hal_context_t *)arg;
  686. portBASE_TYPE HPTaskAwoken = pdFALSE;
  687. // Tx end interrupt
  688. status = rmt_ll_get_tx_end_interrupt_status(hal->regs);
  689. while (status) {
  690. channel = __builtin_ffs(status) - 1;
  691. status &= ~(1 << channel);
  692. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  693. if (p_rmt) {
  694. xSemaphoreGiveFromISR(p_rmt->tx_sem, &HPTaskAwoken);
  695. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  696. p_rmt->tx_data = NULL;
  697. p_rmt->tx_len_rem = 0;
  698. p_rmt->tx_offset = 0;
  699. p_rmt->tx_sub_len = 0;
  700. p_rmt->sample_cur = NULL;
  701. p_rmt->translator = false;
  702. if (rmt_contex.rmt_tx_end_callback.function != NULL) {
  703. rmt_contex.rmt_tx_end_callback.function(channel, rmt_contex.rmt_tx_end_callback.arg);
  704. }
  705. }
  706. rmt_ll_clear_tx_end_interrupt(hal->regs, channel);
  707. }
  708. // Tx thres interrupt
  709. status = rmt_ll_get_tx_thres_interrupt_status(hal->regs);
  710. while (status) {
  711. channel = __builtin_ffs(status) - 1;
  712. status &= ~(1 << channel);
  713. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  714. if (p_rmt) {
  715. if (p_rmt->translator) {
  716. if (p_rmt->sample_size_remain > 0) {
  717. size_t translated_size = 0;
  718. p_rmt->sample_to_rmt((void *)p_rmt->sample_cur,
  719. p_rmt->tx_buf,
  720. p_rmt->sample_size_remain,
  721. p_rmt->tx_sub_len,
  722. &translated_size,
  723. &p_rmt->tx_len_rem);
  724. p_rmt->sample_size_remain -= translated_size;
  725. p_rmt->sample_cur += translated_size;
  726. p_rmt->tx_data = p_rmt->tx_buf;
  727. } else {
  728. p_rmt->sample_cur = NULL;
  729. p_rmt->translator = false;
  730. }
  731. }
  732. const rmt_item32_t *pdata = p_rmt->tx_data;
  733. size_t len_rem = p_rmt->tx_len_rem;
  734. if (len_rem >= p_rmt->tx_sub_len) {
  735. rmt_fill_memory(channel, pdata, p_rmt->tx_sub_len, p_rmt->tx_offset);
  736. p_rmt->tx_data += p_rmt->tx_sub_len;
  737. p_rmt->tx_len_rem -= p_rmt->tx_sub_len;
  738. } else if (len_rem == 0) {
  739. rmt_item32_t stop_data = {0};
  740. rmt_ll_write_memory(rmt_contex.hal.mem, channel, &stop_data, 1, p_rmt->tx_offset);
  741. } else {
  742. rmt_fill_memory(channel, pdata, len_rem, p_rmt->tx_offset);
  743. rmt_item32_t stop_data = {0};
  744. rmt_ll_write_memory(rmt_contex.hal.mem, channel, &stop_data, 1, p_rmt->tx_offset + len_rem);
  745. p_rmt->tx_data += len_rem;
  746. p_rmt->tx_len_rem -= len_rem;
  747. }
  748. if (p_rmt->tx_offset == 0) {
  749. p_rmt->tx_offset = p_rmt->tx_sub_len;
  750. } else {
  751. p_rmt->tx_offset = 0;
  752. }
  753. }
  754. rmt_ll_clear_tx_thres_interrupt(hal->regs, channel);
  755. }
  756. // Rx end interrupt
  757. status = rmt_ll_get_rx_end_interrupt_status(hal->regs);
  758. while (status) {
  759. channel = __builtin_ffs(status) - 1;
  760. status &= ~(1 << channel);
  761. rmt_obj_t *p_rmt = p_rmt_obj[RMT_ENCODE_RX_CHANNEL(channel)];
  762. if (p_rmt) {
  763. rmt_ll_rx_enable(rmt_contex.hal.regs, channel, false);
  764. int item_len = rmt_rx_get_mem_len_in_isr(channel);
  765. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, channel, RMT_MEM_OWNER_SW);
  766. if (p_rmt->rx_buf) {
  767. addr = RMTMEM.chan[RMT_ENCODE_RX_CHANNEL(channel)].data32;
  768. #if SOC_RMT_SUPPORT_RX_PINGPONG
  769. if (item_len > p_rmt->rx_item_start_idx) {
  770. item_len = item_len - p_rmt->rx_item_start_idx;
  771. }
  772. memcpy((void *)(p_rmt->rx_item_buf + p_rmt->rx_item_len), (void *)(addr + p_rmt->rx_item_start_idx), item_len * 4);
  773. p_rmt->rx_item_len += item_len;
  774. BaseType_t res = xRingbufferSendFromISR(p_rmt->rx_buf, (void *)(p_rmt->rx_item_buf), p_rmt->rx_item_len * 4, &HPTaskAwoken);
  775. #else
  776. BaseType_t res = xRingbufferSendFromISR(p_rmt->rx_buf, (void *)addr, item_len * 4, &HPTaskAwoken);
  777. #endif
  778. if (res == pdFALSE) {
  779. ESP_EARLY_LOGE(RMT_TAG, "RMT RX BUFFER FULL");
  780. }
  781. } else {
  782. ESP_EARLY_LOGE(RMT_TAG, "RMT RX BUFFER ERROR");
  783. }
  784. #if SOC_RMT_SUPPORT_RX_PINGPONG
  785. p_rmt->rx_item_start_idx = 0;
  786. p_rmt->rx_item_len = 0;
  787. memset((void *)p_rmt->rx_item_buf, 0, p_rmt->rx_item_buf_size);
  788. #endif
  789. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, channel);
  790. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, channel, RMT_MEM_OWNER_HW);
  791. rmt_ll_rx_enable(rmt_contex.hal.regs, channel, true);
  792. }
  793. rmt_ll_clear_rx_end_interrupt(hal->regs, channel);
  794. }
  795. #if SOC_RMT_SUPPORT_RX_PINGPONG
  796. // Rx thres interrupt
  797. status = rmt_ll_get_rx_thres_interrupt_status(hal->regs);
  798. while (status) {
  799. channel = __builtin_ffs(status) - 1;
  800. status &= ~(1 << channel);
  801. rmt_obj_t *p_rmt = p_rmt_obj[RMT_ENCODE_RX_CHANNEL(channel)];
  802. int mem_item_size = rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, channel) * RMT_MEM_ITEM_NUM;
  803. int rx_thres_lim = rmt_ll_rx_get_limit(rmt_contex.hal.regs, channel);
  804. int item_len = (p_rmt->rx_item_start_idx == 0) ? rx_thres_lim : (mem_item_size - rx_thres_lim);
  805. if ((p_rmt->rx_item_len + item_len) < (p_rmt->rx_item_buf_size / 4)) {
  806. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, channel, RMT_MEM_OWNER_SW);
  807. memcpy((void *)(p_rmt->rx_item_buf + p_rmt->rx_item_len), (void *)(RMTMEM.chan[RMT_ENCODE_RX_CHANNEL(channel)].data32 + p_rmt->rx_item_start_idx), item_len * 4);
  808. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, channel, RMT_MEM_OWNER_HW);
  809. p_rmt->rx_item_len += item_len;
  810. p_rmt->rx_item_start_idx += item_len;
  811. if (p_rmt->rx_item_start_idx >= mem_item_size) {
  812. p_rmt->rx_item_start_idx = 0;
  813. }
  814. } else {
  815. ESP_EARLY_LOGE(RMT_TAG, "---RX buffer too small: %d", sizeof(p_rmt->rx_item_buf));
  816. }
  817. rmt_ll_clear_rx_thres_interrupt(hal->regs, channel);
  818. }
  819. #endif
  820. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  821. // loop count interrupt
  822. status = rmt_ll_get_tx_loop_interrupt_status(hal->regs);
  823. while (status) {
  824. channel = __builtin_ffs(status) - 1;
  825. status &= ~(1 << channel);
  826. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  827. if (p_rmt) {
  828. xSemaphoreGiveFromISR(p_rmt->tx_sem, &HPTaskAwoken);
  829. if (rmt_contex.rmt_tx_end_callback.function != NULL) {
  830. rmt_contex.rmt_tx_end_callback.function(channel, rmt_contex.rmt_tx_end_callback.arg);
  831. }
  832. }
  833. rmt_ll_clear_tx_loop_interrupt(hal->regs, channel);
  834. }
  835. #endif
  836. // RX Err interrupt
  837. status = rmt_ll_get_rx_err_interrupt_status(hal->regs);
  838. while (status) {
  839. channel = __builtin_ffs(status) - 1;
  840. status &= ~(1 << channel);
  841. rmt_obj_t *p_rmt = p_rmt_obj[RMT_ENCODE_RX_CHANNEL(channel)];
  842. if (p_rmt) {
  843. // Reset the receiver's write/read addresses to prevent endless err interrupts.
  844. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, channel);
  845. ESP_EARLY_LOGD(RMT_TAG, "RMT RX channel %d error", channel);
  846. ESP_EARLY_LOGD(RMT_TAG, "status: 0x%08x", rmt_ll_rx_get_channel_status(rmt_contex.hal.regs, channel));
  847. }
  848. rmt_ll_clear_rx_err_interrupt(hal->regs, channel);
  849. }
  850. // TX Err interrupt
  851. status = rmt_ll_get_tx_err_interrupt_status(hal->regs);
  852. while (status) {
  853. channel = __builtin_ffs(status) - 1;
  854. status &= ~(1 << channel);
  855. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  856. if (p_rmt) {
  857. // Reset the transmitter's write/read addresses to prevent endless err interrupts.
  858. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  859. ESP_EARLY_LOGD(RMT_TAG, "RMT TX channel %d error", channel);
  860. ESP_EARLY_LOGD(RMT_TAG, "status: 0x%08x", rmt_ll_tx_get_channel_status(rmt_contex.hal.regs, channel));
  861. }
  862. rmt_ll_clear_tx_err_interrupt(hal->regs, channel);
  863. }
  864. if (HPTaskAwoken == pdTRUE) {
  865. portYIELD_FROM_ISR();
  866. }
  867. }
  868. esp_err_t rmt_driver_uninstall(rmt_channel_t channel)
  869. {
  870. esp_err_t err = ESP_OK;
  871. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  872. RMT_CHECK((rmt_contex.rmt_driver_channels & BIT(channel)) != 0, "No RMT driver for this channel", ESP_ERR_INVALID_STATE);
  873. if (p_rmt_obj[channel] == NULL) {
  874. return ESP_OK;
  875. }
  876. //Avoid blocking here(when the interrupt is disabled and do not wait tx done).
  877. if (p_rmt_obj[channel]->wait_done) {
  878. xSemaphoreTake(p_rmt_obj[channel]->tx_sem, portMAX_DELAY);
  879. }
  880. RMT_ENTER_CRITICAL();
  881. // check channel's working mode
  882. if (p_rmt_obj[channel]->rx_buf) {
  883. rmt_ll_enable_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), 0);
  884. rmt_ll_enable_rx_err_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), 0);
  885. #if SOC_RMT_SUPPORT_RX_PINGPONG
  886. rmt_ll_enable_rx_thres_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), 0);
  887. #endif
  888. } else {
  889. rmt_ll_enable_tx_end_interrupt(rmt_contex.hal.regs, channel, 0);
  890. rmt_ll_enable_tx_err_interrupt(rmt_contex.hal.regs, channel, 0);
  891. rmt_ll_enable_tx_thres_interrupt(rmt_contex.hal.regs, channel, false);
  892. }
  893. RMT_EXIT_CRITICAL();
  894. _lock_acquire_recursive(&(rmt_contex.rmt_driver_isr_lock));
  895. rmt_contex.rmt_driver_channels &= ~BIT(channel);
  896. if (rmt_contex.rmt_driver_channels == 0) {
  897. rmt_module_disable();
  898. // all channels have driver disabled
  899. err = rmt_isr_deregister(rmt_contex.rmt_driver_intr_handle);
  900. rmt_contex.rmt_driver_intr_handle = NULL;
  901. }
  902. _lock_release_recursive(&(rmt_contex.rmt_driver_isr_lock));
  903. if (err != ESP_OK) {
  904. return err;
  905. }
  906. if (p_rmt_obj[channel]->tx_sem) {
  907. vSemaphoreDelete(p_rmt_obj[channel]->tx_sem);
  908. p_rmt_obj[channel]->tx_sem = NULL;
  909. }
  910. if (p_rmt_obj[channel]->rx_buf) {
  911. vRingbufferDelete(p_rmt_obj[channel]->rx_buf);
  912. p_rmt_obj[channel]->rx_buf = NULL;
  913. }
  914. if (p_rmt_obj[channel]->tx_buf) {
  915. free(p_rmt_obj[channel]->tx_buf);
  916. p_rmt_obj[channel]->tx_buf = NULL;
  917. }
  918. if (p_rmt_obj[channel]->sample_to_rmt) {
  919. p_rmt_obj[channel]->sample_to_rmt = NULL;
  920. }
  921. #if SOC_RMT_SUPPORT_RX_PINGPONG
  922. if (p_rmt_obj[channel]->rx_item_buf) {
  923. free(p_rmt_obj[channel]->rx_item_buf);
  924. p_rmt_obj[channel]->rx_item_buf = NULL;
  925. p_rmt_obj[channel]->rx_item_buf_size = 0;
  926. }
  927. #endif
  928. free(p_rmt_obj[channel]);
  929. p_rmt_obj[channel] = NULL;
  930. return ESP_OK;
  931. }
  932. esp_err_t rmt_driver_install(rmt_channel_t channel, size_t rx_buf_size, int intr_alloc_flags)
  933. {
  934. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  935. RMT_CHECK((rmt_contex.rmt_driver_channels & BIT(channel)) == 0,
  936. "RMT driver already installed for channel", ESP_ERR_INVALID_STATE);
  937. esp_err_t err = ESP_OK;
  938. if (p_rmt_obj[channel] != NULL) {
  939. ESP_LOGD(RMT_TAG, "RMT driver already installed");
  940. return ESP_ERR_INVALID_STATE;
  941. }
  942. #if !CONFIG_SPIRAM_USE_MALLOC
  943. p_rmt_obj[channel] = calloc(1, sizeof(rmt_obj_t));
  944. #else
  945. if (!(intr_alloc_flags & ESP_INTR_FLAG_IRAM)) {
  946. p_rmt_obj[channel] = calloc(1, sizeof(rmt_obj_t));
  947. } else {
  948. p_rmt_obj[channel] = heap_caps_calloc(1, sizeof(rmt_obj_t), MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT);
  949. }
  950. #endif
  951. if (p_rmt_obj[channel] == NULL) {
  952. ESP_LOGE(RMT_TAG, "RMT driver malloc error");
  953. return ESP_ERR_NO_MEM;
  954. }
  955. p_rmt_obj[channel]->tx_len_rem = 0;
  956. p_rmt_obj[channel]->tx_data = NULL;
  957. p_rmt_obj[channel]->channel = channel;
  958. p_rmt_obj[channel]->tx_offset = 0;
  959. p_rmt_obj[channel]->tx_sub_len = 0;
  960. p_rmt_obj[channel]->wait_done = false;
  961. p_rmt_obj[channel]->translator = false;
  962. p_rmt_obj[channel]->sample_to_rmt = NULL;
  963. if (p_rmt_obj[channel]->tx_sem == NULL) {
  964. #if !CONFIG_SPIRAM_USE_MALLOC
  965. p_rmt_obj[channel]->tx_sem = xSemaphoreCreateBinary();
  966. #else
  967. p_rmt_obj[channel]->intr_alloc_flags = intr_alloc_flags;
  968. if (!(intr_alloc_flags & ESP_INTR_FLAG_IRAM)) {
  969. p_rmt_obj[channel]->tx_sem = xSemaphoreCreateBinary();
  970. } else {
  971. p_rmt_obj[channel]->tx_sem = xSemaphoreCreateBinaryStatic(&p_rmt_obj[channel]->tx_sem_buffer);
  972. }
  973. #endif
  974. xSemaphoreGive(p_rmt_obj[channel]->tx_sem);
  975. }
  976. if (p_rmt_obj[channel]->rx_buf == NULL && rx_buf_size > 0) {
  977. p_rmt_obj[channel]->rx_buf = xRingbufferCreate(rx_buf_size, RINGBUF_TYPE_NOSPLIT);
  978. }
  979. #if SOC_RMT_SUPPORT_RX_PINGPONG
  980. if (p_rmt_obj[channel]->rx_item_buf == NULL && rx_buf_size > 0) {
  981. #if !CONFIG_SPIRAM_USE_MALLOC
  982. p_rmt_obj[channel]->rx_item_buf = calloc(1, rx_buf_size);
  983. #else
  984. if (!(p_rmt_obj[channel]->intr_alloc_flags & ESP_INTR_FLAG_IRAM)) {
  985. p_rmt_obj[channel]->rx_item_buf = calloc(1, rx_buf_size);
  986. } else {
  987. p_rmt_obj[channel]->rx_item_buf = heap_caps_calloc(1, rx_buf_size, MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT);
  988. }
  989. #endif
  990. if (p_rmt_obj[channel]->rx_item_buf == NULL) {
  991. ESP_LOGE(RMT_TAG, "RMT malloc fail");
  992. return ESP_FAIL;
  993. }
  994. p_rmt_obj[channel]->rx_item_buf_size = rx_buf_size;
  995. }
  996. #endif
  997. _lock_acquire_recursive(&(rmt_contex.rmt_driver_isr_lock));
  998. if (rmt_contex.rmt_driver_channels == 0) {
  999. // first RMT channel using driver
  1000. err = rmt_isr_register(rmt_driver_isr_default, &rmt_contex.hal, intr_alloc_flags, &(rmt_contex.rmt_driver_intr_handle));
  1001. }
  1002. if (err == ESP_OK) {
  1003. rmt_contex.rmt_driver_channels |= BIT(channel);
  1004. }
  1005. _lock_release_recursive(&(rmt_contex.rmt_driver_isr_lock));
  1006. rmt_module_enable();
  1007. if (RMT_IS_RX_CHANNEL(channel)) {
  1008. rmt_hal_rx_channel_reset(&rmt_contex.hal, RMT_DECODE_RX_CHANNEL(channel));
  1009. } else {
  1010. rmt_hal_tx_channel_reset(&rmt_contex.hal, channel);
  1011. }
  1012. return err;
  1013. }
  1014. esp_err_t rmt_write_items(rmt_channel_t channel, const rmt_item32_t *rmt_item, int item_num, bool wait_tx_done)
  1015. {
  1016. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1017. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1018. RMT_CHECK(rmt_item != NULL, RMT_ADDR_ERROR_STR, ESP_FAIL);
  1019. RMT_CHECK(item_num > 0, RMT_DRIVER_LENGTH_ERROR_STR, ESP_ERR_INVALID_ARG);
  1020. #if CONFIG_SPIRAM_USE_MALLOC
  1021. if (p_rmt_obj[channel]->intr_alloc_flags & ESP_INTR_FLAG_IRAM) {
  1022. if (!esp_ptr_internal(rmt_item)) {
  1023. ESP_LOGE(RMT_TAG, RMT_PSRAM_BUFFER_WARN_STR);
  1024. return ESP_ERR_INVALID_ARG;
  1025. }
  1026. }
  1027. #endif
  1028. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  1029. int block_num = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel);
  1030. int item_block_len = block_num * RMT_MEM_ITEM_NUM;
  1031. int item_sub_len = block_num * RMT_MEM_ITEM_NUM / 2;
  1032. int len_rem = item_num;
  1033. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1034. // fill the memory block first
  1035. if (item_num >= item_block_len) {
  1036. rmt_fill_memory(channel, rmt_item, item_block_len, 0);
  1037. len_rem -= item_block_len;
  1038. rmt_set_tx_loop_mode(channel, false);
  1039. rmt_set_tx_thr_intr_en(channel, 1, item_sub_len);
  1040. p_rmt->tx_data = rmt_item + item_block_len;
  1041. p_rmt->tx_len_rem = len_rem;
  1042. p_rmt->tx_offset = 0;
  1043. p_rmt->tx_sub_len = item_sub_len;
  1044. } else {
  1045. rmt_fill_memory(channel, rmt_item, len_rem, 0);
  1046. rmt_item32_t stop_data = {0};
  1047. rmt_ll_write_memory(rmt_contex.hal.mem, channel, &stop_data, 1, len_rem);
  1048. p_rmt->tx_len_rem = 0;
  1049. }
  1050. rmt_tx_start(channel, true);
  1051. p_rmt->wait_done = wait_tx_done;
  1052. if (wait_tx_done) {
  1053. // wait loop done
  1054. if (rmt_ll_is_tx_loop_enabled(rmt_contex.hal.regs, channel)) {
  1055. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  1056. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1057. xSemaphoreGive(p_rmt->tx_sem);
  1058. #endif
  1059. } else {
  1060. // wait tx end
  1061. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1062. xSemaphoreGive(p_rmt->tx_sem);
  1063. }
  1064. }
  1065. return ESP_OK;
  1066. }
  1067. esp_err_t rmt_wait_tx_done(rmt_channel_t channel, TickType_t wait_time)
  1068. {
  1069. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1070. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1071. if (xSemaphoreTake(p_rmt_obj[channel]->tx_sem, wait_time) == pdTRUE) {
  1072. p_rmt_obj[channel]->wait_done = false;
  1073. xSemaphoreGive(p_rmt_obj[channel]->tx_sem);
  1074. return ESP_OK;
  1075. } else {
  1076. if (wait_time != 0) {
  1077. // Don't emit error message if just polling.
  1078. ESP_LOGE(RMT_TAG, "Timeout on wait_tx_done");
  1079. }
  1080. return ESP_ERR_TIMEOUT;
  1081. }
  1082. }
  1083. esp_err_t rmt_get_ringbuf_handle(rmt_channel_t channel, RingbufHandle_t *buf_handle)
  1084. {
  1085. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1086. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1087. RMT_CHECK(buf_handle != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  1088. *buf_handle = p_rmt_obj[channel]->rx_buf;
  1089. return ESP_OK;
  1090. }
  1091. rmt_tx_end_callback_t rmt_register_tx_end_callback(rmt_tx_end_fn_t function, void *arg)
  1092. {
  1093. rmt_tx_end_callback_t previous = rmt_contex.rmt_tx_end_callback;
  1094. rmt_contex.rmt_tx_end_callback.function = function;
  1095. rmt_contex.rmt_tx_end_callback.arg = arg;
  1096. return previous;
  1097. }
  1098. esp_err_t rmt_translator_init(rmt_channel_t channel, sample_to_rmt_t fn)
  1099. {
  1100. RMT_CHECK(fn != NULL, RMT_TRANSLATOR_NULL_STR, ESP_ERR_INVALID_ARG);
  1101. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1102. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1103. const uint32_t block_size = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel) *
  1104. RMT_MEM_ITEM_NUM * sizeof(rmt_item32_t);
  1105. if (p_rmt_obj[channel]->tx_buf == NULL) {
  1106. #if !CONFIG_SPIRAM_USE_MALLOC
  1107. p_rmt_obj[channel]->tx_buf = (rmt_item32_t *)malloc(block_size);
  1108. #else
  1109. if (p_rmt_obj[channel]->intr_alloc_flags & ESP_INTR_FLAG_IRAM) {
  1110. p_rmt_obj[channel]->tx_buf = (rmt_item32_t *)malloc(block_size);
  1111. } else {
  1112. p_rmt_obj[channel]->tx_buf = (rmt_item32_t *)heap_caps_calloc(1, block_size, MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT);
  1113. }
  1114. #endif
  1115. if (p_rmt_obj[channel]->tx_buf == NULL) {
  1116. ESP_LOGE(RMT_TAG, "RMT translator buffer create fail");
  1117. return ESP_FAIL;
  1118. }
  1119. }
  1120. p_rmt_obj[channel]->sample_to_rmt = fn;
  1121. p_rmt_obj[channel]->tx_context = NULL;
  1122. p_rmt_obj[channel]->sample_size_remain = 0;
  1123. p_rmt_obj[channel]->sample_cur = NULL;
  1124. ESP_LOGD(RMT_TAG, "RMT translator init done");
  1125. return ESP_OK;
  1126. }
  1127. esp_err_t rmt_translator_set_context(rmt_channel_t channel, void *context)
  1128. {
  1129. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1130. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1131. p_rmt_obj[channel]->tx_context = context;
  1132. return ESP_OK;
  1133. }
  1134. esp_err_t rmt_translator_get_context(const size_t *item_num, void **context)
  1135. {
  1136. RMT_CHECK(item_num && context, "invalid arguments", ESP_ERR_INVALID_ARG);
  1137. // the address of tx_len_rem is directlly passed to the callback,
  1138. // so it's possible to get the object address from that
  1139. rmt_obj_t *obj = __containerof(item_num, rmt_obj_t, tx_len_rem);
  1140. *context = obj->tx_context;
  1141. return ESP_OK;
  1142. }
  1143. esp_err_t rmt_write_sample(rmt_channel_t channel, const uint8_t *src, size_t src_size, bool wait_tx_done)
  1144. {
  1145. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1146. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1147. RMT_CHECK(p_rmt_obj[channel]->sample_to_rmt != NULL, RMT_TRANSLATOR_UNINIT_STR, ESP_FAIL);
  1148. #if CONFIG_SPIRAM_USE_MALLOC
  1149. if (p_rmt_obj[channel]->intr_alloc_flags & ESP_INTR_FLAG_IRAM) {
  1150. if (!esp_ptr_internal(src)) {
  1151. ESP_LOGE(RMT_TAG, RMT_PSRAM_BUFFER_WARN_STR);
  1152. return ESP_ERR_INVALID_ARG;
  1153. }
  1154. }
  1155. #endif
  1156. size_t translated_size = 0;
  1157. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  1158. const uint32_t item_block_len = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel) * RMT_MEM_ITEM_NUM;
  1159. const uint32_t item_sub_len = item_block_len / 2;
  1160. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1161. p_rmt->sample_to_rmt((void *)src, p_rmt->tx_buf, src_size, item_block_len, &translated_size, &p_rmt->tx_len_rem);
  1162. p_rmt->sample_size_remain = src_size - translated_size;
  1163. p_rmt->sample_cur = src + translated_size;
  1164. rmt_fill_memory(channel, p_rmt->tx_buf, p_rmt->tx_len_rem, 0);
  1165. if (p_rmt->tx_len_rem == item_block_len) {
  1166. rmt_set_tx_thr_intr_en(channel, 1, item_sub_len);
  1167. p_rmt->tx_data = p_rmt->tx_buf;
  1168. p_rmt->tx_offset = 0;
  1169. p_rmt->tx_sub_len = item_sub_len;
  1170. p_rmt->translator = true;
  1171. } else {
  1172. rmt_item32_t stop_data = {0};
  1173. rmt_ll_write_memory(rmt_contex.hal.mem, channel, &stop_data, 1, p_rmt->tx_len_rem);
  1174. p_rmt->tx_len_rem = 0;
  1175. p_rmt->sample_cur = NULL;
  1176. p_rmt->translator = false;
  1177. }
  1178. rmt_tx_start(channel, true);
  1179. p_rmt->wait_done = wait_tx_done;
  1180. if (wait_tx_done) {
  1181. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1182. xSemaphoreGive(p_rmt->tx_sem);
  1183. }
  1184. return ESP_OK;
  1185. }
  1186. esp_err_t rmt_get_channel_status(rmt_channel_status_result_t *channel_status)
  1187. {
  1188. RMT_CHECK(channel_status != NULL, RMT_PARAM_ERR_STR, ESP_ERR_INVALID_ARG);
  1189. for (int i = 0; i < RMT_CHANNEL_MAX; i++) {
  1190. channel_status->status[i] = RMT_CHANNEL_UNINIT;
  1191. if (p_rmt_obj[i] != NULL) {
  1192. if (p_rmt_obj[i]->tx_sem != NULL) {
  1193. if (xSemaphoreTake(p_rmt_obj[i]->tx_sem, (TickType_t)0) == pdTRUE) {
  1194. channel_status->status[i] = RMT_CHANNEL_IDLE;
  1195. xSemaphoreGive(p_rmt_obj[i]->tx_sem);
  1196. } else {
  1197. channel_status->status[i] = RMT_CHANNEL_BUSY;
  1198. }
  1199. }
  1200. }
  1201. }
  1202. return ESP_OK;
  1203. }
  1204. esp_err_t rmt_get_counter_clock(rmt_channel_t channel, uint32_t *clock_hz)
  1205. {
  1206. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1207. RMT_CHECK(clock_hz, "parameter clock_hz can't be null", ESP_ERR_INVALID_ARG);
  1208. RMT_ENTER_CRITICAL();
  1209. uint32_t rmt_source_clk_hz = 0;
  1210. #if SOC_RMT_SOURCE_CLK_INDEPENDENT
  1211. rmt_source_clk_hz = s_rmt_source_clock_hz[channel];
  1212. #else
  1213. rmt_source_clk_hz = s_rmt_source_clock_hz;
  1214. #endif
  1215. if (RMT_IS_RX_CHANNEL(channel)) {
  1216. *clock_hz = rmt_source_clk_hz / rmt_ll_rx_get_channel_clock_div(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  1217. } else {
  1218. *clock_hz = rmt_source_clk_hz / rmt_ll_tx_get_channel_clock_div(rmt_contex.hal.regs, channel);
  1219. }
  1220. RMT_EXIT_CRITICAL();
  1221. return ESP_OK;
  1222. }
  1223. #if SOC_RMT_SUPPORT_TX_SYNCHRO
  1224. esp_err_t rmt_add_channel_to_group(rmt_channel_t channel)
  1225. {
  1226. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1227. RMT_ENTER_CRITICAL();
  1228. rmt_ll_tx_enable_sync(rmt_contex.hal.regs, true);
  1229. rmt_contex.synchro_channel_mask |= (1 << channel);
  1230. rmt_ll_tx_add_to_sync_group(rmt_contex.hal.regs, channel);
  1231. rmt_ll_tx_reset_channels_clock_div(rmt_contex.hal.regs, rmt_contex.synchro_channel_mask);
  1232. RMT_EXIT_CRITICAL();
  1233. return ESP_OK;
  1234. }
  1235. esp_err_t rmt_remove_channel_from_group(rmt_channel_t channel)
  1236. {
  1237. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1238. RMT_ENTER_CRITICAL();
  1239. rmt_contex.synchro_channel_mask &= ~(1 << channel);
  1240. rmt_ll_tx_remove_from_sync_group(rmt_contex.hal.regs, channel);
  1241. if (rmt_contex.synchro_channel_mask == 0) {
  1242. rmt_ll_tx_enable_sync(rmt_contex.hal.regs, false);
  1243. }
  1244. RMT_EXIT_CRITICAL();
  1245. return ESP_OK;
  1246. }
  1247. esp_err_t rmt_memory_rw_rst(rmt_channel_t channel)
  1248. {
  1249. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1250. RMT_ENTER_CRITICAL();
  1251. if (RMT_IS_RX_CHANNEL(channel)) {
  1252. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  1253. } else {
  1254. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  1255. }
  1256. RMT_EXIT_CRITICAL();
  1257. return ESP_OK;
  1258. }
  1259. #endif
  1260. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  1261. esp_err_t rmt_set_tx_loop_count(rmt_channel_t channel, uint32_t count)
  1262. {
  1263. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1264. RMT_ENTER_CRITICAL();
  1265. rmt_ll_tx_set_loop_count(rmt_contex.hal.regs, channel, count);
  1266. RMT_EXIT_CRITICAL();
  1267. return ESP_OK;
  1268. }
  1269. #endif