rmt.c 54 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379
  1. // Copyright 2015-2020 Espressif Systems (Shanghai) PTE LTD
  2. //
  3. // Licensed under the Apache License, Version 2.0 (the "License");
  4. // you may not use this file except in compliance with the License.
  5. // You may obtain a copy of the License at
  6. //
  7. // http://www.apache.org/licenses/LICENSE-2.0
  8. //
  9. // Unless required by applicable law or agreed to in writing, software
  10. // distributed under the License is distributed on an "AS IS" BASIS,
  11. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. // See the License for the specific language governing permissions and
  13. // limitations under the License.
  14. #include <stdlib.h>
  15. #include <string.h>
  16. #include <sys/lock.h>
  17. #include <sys/cdefs.h>
  18. #include "esp_compiler.h"
  19. #include "esp_intr_alloc.h"
  20. #include "esp_log.h"
  21. #include "driver/gpio.h"
  22. #include "driver/periph_ctrl.h"
  23. #include "driver/rmt.h"
  24. #include "freertos/FreeRTOS.h"
  25. #include "freertos/task.h"
  26. #include "freertos/semphr.h"
  27. #include "freertos/ringbuf.h"
  28. #include "soc/soc_memory_layout.h"
  29. #include "soc/rmt_periph.h"
  30. #include "soc/rtc.h"
  31. #include "hal/rmt_hal.h"
  32. #include "hal/rmt_ll.h"
  33. #include "hal/gpio_hal.h"
  34. #include "esp_rom_gpio.h"
  35. #define RMT_CHANNEL_ERROR_STR "RMT CHANNEL ERR"
  36. #define RMT_ADDR_ERROR_STR "RMT ADDRESS ERR"
  37. #define RMT_MEM_CNT_ERROR_STR "RMT MEM BLOCK NUM ERR"
  38. #define RMT_CARRIER_ERROR_STR "RMT CARRIER LEVEL ERR"
  39. #define RMT_MEM_OWNER_ERROR_STR "RMT MEM OWNER_ERR"
  40. #define RMT_BASECLK_ERROR_STR "RMT BASECLK ERR"
  41. #define RMT_WR_MEM_OVF_ERROR_STR "RMT WR MEM OVERFLOW"
  42. #define RMT_GPIO_ERROR_STR "RMT GPIO ERROR"
  43. #define RMT_MODE_ERROR_STR "RMT MODE ERROR"
  44. #define RMT_CLK_DIV_ERROR_STR "RMT CLK DIV ERR"
  45. #define RMT_DRIVER_ERROR_STR "RMT DRIVER ERR"
  46. #define RMT_DRIVER_LENGTH_ERROR_STR "RMT PARAM LEN ERROR"
  47. #define RMT_PSRAM_BUFFER_WARN_STR "Using buffer allocated from psram"
  48. #define RMT_TRANSLATOR_NULL_STR "RMT translator is null"
  49. #define RMT_TRANSLATOR_UNINIT_STR "RMT translator not init"
  50. #define RMT_PARAM_ERR_STR "RMT param error"
  51. static const char *RMT_TAG = "rmt";
  52. #define RMT_CHECK(a, str, ret_val, ...) \
  53. if (unlikely(!(a))) { \
  54. ESP_LOGE(RMT_TAG, "%s(%d): "str, __FUNCTION__, __LINE__, ##__VA_ARGS__); \
  55. return (ret_val); \
  56. }
  57. // Spinlock for protecting concurrent register-level access only
  58. #define RMT_ENTER_CRITICAL() portENTER_CRITICAL_SAFE(&(rmt_contex.rmt_spinlock))
  59. #define RMT_EXIT_CRITICAL() portEXIT_CRITICAL_SAFE(&(rmt_contex.rmt_spinlock))
  60. #define RMT_RX_CHANNEL_ENCODING_START (SOC_RMT_CHANNELS_NUM-SOC_RMT_TX_CHANNELS_NUM)
  61. #define RMT_TX_CHANNEL_ENCODING_END (SOC_RMT_TX_CHANNELS_NUM-1)
  62. #define RMT_IS_RX_CHANNEL(channel) ((channel) >= RMT_RX_CHANNEL_ENCODING_START)
  63. #define RMT_IS_TX_CHANNEL(channel) ((channel) <= RMT_TX_CHANNEL_ENCODING_END)
  64. #define RMT_DECODE_RX_CHANNEL(encode_chan) ((encode_chan - RMT_RX_CHANNEL_ENCODING_START))
  65. #define RMT_ENCODE_RX_CHANNEL(decode_chan) ((decode_chan + RMT_RX_CHANNEL_ENCODING_START))
  66. typedef struct {
  67. rmt_hal_context_t hal;
  68. _lock_t rmt_driver_isr_lock;
  69. portMUX_TYPE rmt_spinlock; // Mutex lock for protecting concurrent register/unregister of RMT channels' ISR
  70. rmt_isr_handle_t rmt_driver_intr_handle;
  71. rmt_tx_end_callback_t rmt_tx_end_callback;// Event called when transmission is ended
  72. uint8_t rmt_driver_channels; // Bitmask of installed drivers' channels
  73. bool rmt_module_enabled;
  74. } rmt_contex_t;
  75. typedef struct {
  76. size_t tx_offset;
  77. size_t tx_len_rem;
  78. size_t tx_sub_len;
  79. bool translator;
  80. bool wait_done; //Mark whether wait tx done.
  81. rmt_channel_t channel;
  82. const rmt_item32_t *tx_data;
  83. xSemaphoreHandle tx_sem;
  84. #if CONFIG_SPIRAM_USE_MALLOC
  85. int intr_alloc_flags;
  86. StaticSemaphore_t tx_sem_buffer;
  87. #endif
  88. rmt_item32_t *tx_buf;
  89. RingbufHandle_t rx_buf;
  90. #if SOC_RMT_SUPPORT_RX_PINGPONG
  91. rmt_item32_t *rx_item_buf;
  92. uint32_t rx_item_buf_size;
  93. uint32_t rx_item_len;
  94. int rx_item_start_idx;
  95. #endif
  96. sample_to_rmt_t sample_to_rmt;
  97. void *tx_context;
  98. size_t sample_size_remain;
  99. const uint8_t *sample_cur;
  100. } rmt_obj_t;
  101. static rmt_contex_t rmt_contex = {
  102. .hal.regs = RMT_LL_HW_BASE,
  103. .hal.mem = RMT_LL_MEM_BASE,
  104. .rmt_spinlock = portMUX_INITIALIZER_UNLOCKED,
  105. .rmt_driver_intr_handle = NULL,
  106. .rmt_tx_end_callback = {
  107. .function = NULL,
  108. },
  109. .rmt_driver_channels = 0,
  110. .rmt_module_enabled = false,
  111. };
  112. static rmt_obj_t *p_rmt_obj[RMT_CHANNEL_MAX] = {0};
  113. #if SOC_RMT_SOURCE_CLK_INDEPENDENT
  114. static uint32_t s_rmt_source_clock_hz[RMT_CHANNEL_MAX];
  115. #else
  116. static uint32_t s_rmt_source_clock_hz;
  117. #endif
  118. //Enable RMT module
  119. static void rmt_module_enable(void)
  120. {
  121. RMT_ENTER_CRITICAL();
  122. if (rmt_contex.rmt_module_enabled == false) {
  123. periph_module_reset(rmt_periph_signals.module);
  124. periph_module_enable(rmt_periph_signals.module);
  125. rmt_contex.rmt_module_enabled = true;
  126. }
  127. RMT_EXIT_CRITICAL();
  128. }
  129. //Disable RMT module
  130. static void rmt_module_disable(void)
  131. {
  132. RMT_ENTER_CRITICAL();
  133. if (rmt_contex.rmt_module_enabled == true) {
  134. periph_module_disable(rmt_periph_signals.module);
  135. rmt_contex.rmt_module_enabled = false;
  136. }
  137. RMT_EXIT_CRITICAL();
  138. }
  139. esp_err_t rmt_set_clk_div(rmt_channel_t channel, uint8_t div_cnt)
  140. {
  141. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  142. RMT_ENTER_CRITICAL();
  143. if (RMT_IS_RX_CHANNEL(channel)) {
  144. rmt_ll_rx_set_counter_clock_div(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), div_cnt);
  145. } else {
  146. rmt_ll_tx_set_counter_clock_div(rmt_contex.hal.regs, channel, div_cnt);
  147. }
  148. RMT_EXIT_CRITICAL();
  149. return ESP_OK;
  150. }
  151. esp_err_t rmt_get_clk_div(rmt_channel_t channel, uint8_t *div_cnt)
  152. {
  153. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  154. RMT_CHECK(div_cnt != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  155. RMT_ENTER_CRITICAL();
  156. if (RMT_IS_RX_CHANNEL(channel)) {
  157. *div_cnt = (uint8_t)rmt_ll_rx_get_counter_clock_div(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  158. } else {
  159. *div_cnt = (uint8_t)rmt_ll_tx_get_counter_clock_div(rmt_contex.hal.regs, channel);
  160. }
  161. RMT_EXIT_CRITICAL();
  162. return ESP_OK;
  163. }
  164. esp_err_t rmt_set_rx_idle_thresh(rmt_channel_t channel, uint16_t thresh)
  165. {
  166. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  167. RMT_ENTER_CRITICAL();
  168. rmt_ll_rx_set_idle_thres(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), thresh);
  169. RMT_EXIT_CRITICAL();
  170. return ESP_OK;
  171. }
  172. esp_err_t rmt_get_rx_idle_thresh(rmt_channel_t channel, uint16_t *thresh)
  173. {
  174. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  175. RMT_CHECK(thresh != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  176. RMT_ENTER_CRITICAL();
  177. *thresh = (uint16_t)rmt_ll_rx_get_idle_thres(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  178. RMT_EXIT_CRITICAL();
  179. return ESP_OK;
  180. }
  181. esp_err_t rmt_set_mem_block_num(rmt_channel_t channel, uint8_t rmt_mem_num)
  182. {
  183. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  184. RMT_CHECK(rmt_mem_num <= RMT_CHANNEL_MAX - channel, RMT_MEM_CNT_ERROR_STR, ESP_ERR_INVALID_ARG);
  185. RMT_ENTER_CRITICAL();
  186. if (RMT_IS_RX_CHANNEL(channel)) {
  187. rmt_ll_rx_set_mem_blocks(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), rmt_mem_num);
  188. } else {
  189. rmt_ll_tx_set_mem_blocks(rmt_contex.hal.regs, channel, rmt_mem_num);
  190. }
  191. RMT_EXIT_CRITICAL();
  192. return ESP_OK;
  193. }
  194. esp_err_t rmt_get_mem_block_num(rmt_channel_t channel, uint8_t *rmt_mem_num)
  195. {
  196. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  197. RMT_CHECK(rmt_mem_num != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  198. RMT_ENTER_CRITICAL();
  199. if (RMT_IS_RX_CHANNEL(channel)) {
  200. *rmt_mem_num = (uint8_t)rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  201. } else {
  202. *rmt_mem_num = (uint8_t)rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel);
  203. }
  204. RMT_EXIT_CRITICAL();
  205. return ESP_OK;
  206. }
  207. esp_err_t rmt_set_tx_carrier(rmt_channel_t channel, bool carrier_en, uint16_t high_level, uint16_t low_level,
  208. rmt_carrier_level_t carrier_level)
  209. {
  210. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  211. RMT_CHECK(carrier_level < RMT_CARRIER_LEVEL_MAX, RMT_CARRIER_ERROR_STR, ESP_ERR_INVALID_ARG);
  212. RMT_ENTER_CRITICAL();
  213. rmt_ll_tx_set_carrier_high_low_ticks(rmt_contex.hal.regs, channel, high_level, low_level);
  214. rmt_ll_tx_set_carrier_level(rmt_contex.hal.regs, channel, carrier_level);
  215. rmt_ll_tx_enable_carrier_modulation(rmt_contex.hal.regs, channel, carrier_en);
  216. RMT_EXIT_CRITICAL();
  217. return ESP_OK;
  218. }
  219. esp_err_t rmt_set_mem_pd(rmt_channel_t channel, bool pd_en)
  220. {
  221. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  222. RMT_ENTER_CRITICAL();
  223. rmt_ll_power_down_mem(rmt_contex.hal.regs, pd_en);
  224. RMT_EXIT_CRITICAL();
  225. return ESP_OK;
  226. }
  227. esp_err_t rmt_get_mem_pd(rmt_channel_t channel, bool *pd_en)
  228. {
  229. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  230. RMT_ENTER_CRITICAL();
  231. *pd_en = rmt_ll_is_mem_power_down(rmt_contex.hal.regs);
  232. RMT_EXIT_CRITICAL();
  233. return ESP_OK;
  234. }
  235. esp_err_t rmt_tx_start(rmt_channel_t channel, bool tx_idx_rst)
  236. {
  237. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  238. RMT_ENTER_CRITICAL();
  239. if (tx_idx_rst) {
  240. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  241. }
  242. rmt_ll_clear_tx_end_interrupt(rmt_contex.hal.regs, channel);
  243. // enable tx end interrupt in non-loop mode
  244. if (!rmt_ll_is_tx_loop_enabled(rmt_contex.hal.regs, channel)) {
  245. rmt_ll_enable_tx_end_interrupt(rmt_contex.hal.regs, channel, true);
  246. } else {
  247. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  248. rmt_ll_tx_reset_loop(rmt_contex.hal.regs, channel);
  249. rmt_ll_tx_enable_loop_count(rmt_contex.hal.regs, channel, true);
  250. rmt_ll_clear_tx_loop_interrupt(rmt_contex.hal.regs, channel);
  251. rmt_ll_enable_tx_loop_interrupt(rmt_contex.hal.regs, channel, true);
  252. #endif
  253. }
  254. rmt_ll_tx_start(rmt_contex.hal.regs, channel);
  255. RMT_EXIT_CRITICAL();
  256. return ESP_OK;
  257. }
  258. esp_err_t rmt_tx_stop(rmt_channel_t channel)
  259. {
  260. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  261. RMT_ENTER_CRITICAL();
  262. rmt_ll_tx_stop(rmt_contex.hal.regs, channel);
  263. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  264. RMT_EXIT_CRITICAL();
  265. return ESP_OK;
  266. }
  267. esp_err_t rmt_rx_start(rmt_channel_t channel, bool rx_idx_rst)
  268. {
  269. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  270. RMT_ENTER_CRITICAL();
  271. rmt_ll_rx_enable(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  272. if (rx_idx_rst) {
  273. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  274. }
  275. rmt_ll_clear_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  276. rmt_ll_enable_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), true);
  277. #if SOC_RMT_SUPPORT_RX_PINGPONG
  278. const uint32_t item_block_len = rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel)) * RMT_MEM_ITEM_NUM;
  279. p_rmt_obj[channel]->rx_item_start_idx = 0;
  280. p_rmt_obj[channel]->rx_item_len = 0;
  281. rmt_set_rx_thr_intr_en(channel, true, item_block_len / 2);
  282. #endif
  283. rmt_ll_rx_enable(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), true);
  284. RMT_EXIT_CRITICAL();
  285. return ESP_OK;
  286. }
  287. esp_err_t rmt_rx_stop(rmt_channel_t channel)
  288. {
  289. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  290. RMT_ENTER_CRITICAL();
  291. rmt_ll_enable_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  292. rmt_ll_rx_enable(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  293. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  294. #if SOC_RMT_SUPPORT_RX_PINGPONG
  295. rmt_ll_enable_rx_thres_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  296. #endif
  297. RMT_EXIT_CRITICAL();
  298. return ESP_OK;
  299. }
  300. esp_err_t rmt_tx_memory_reset(rmt_channel_t channel)
  301. {
  302. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  303. RMT_ENTER_CRITICAL();
  304. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  305. RMT_EXIT_CRITICAL();
  306. return ESP_OK;
  307. }
  308. esp_err_t rmt_rx_memory_reset(rmt_channel_t channel)
  309. {
  310. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  311. RMT_ENTER_CRITICAL();
  312. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  313. RMT_EXIT_CRITICAL();
  314. return ESP_OK;
  315. }
  316. esp_err_t rmt_set_memory_owner(rmt_channel_t channel, rmt_mem_owner_t owner)
  317. {
  318. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  319. RMT_CHECK(owner < RMT_MEM_OWNER_MAX, RMT_MEM_OWNER_ERROR_STR, ESP_ERR_INVALID_ARG);
  320. RMT_ENTER_CRITICAL();
  321. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), owner);
  322. RMT_EXIT_CRITICAL();
  323. return ESP_OK;
  324. }
  325. esp_err_t rmt_get_memory_owner(rmt_channel_t channel, rmt_mem_owner_t *owner)
  326. {
  327. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  328. RMT_CHECK(owner != NULL, RMT_MEM_OWNER_ERROR_STR, ESP_ERR_INVALID_ARG);
  329. RMT_ENTER_CRITICAL();
  330. *owner = (rmt_mem_owner_t)rmt_ll_rx_get_mem_owner(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  331. RMT_EXIT_CRITICAL();
  332. return ESP_OK;
  333. }
  334. esp_err_t rmt_set_tx_loop_mode(rmt_channel_t channel, bool loop_en)
  335. {
  336. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  337. RMT_ENTER_CRITICAL();
  338. rmt_ll_tx_enable_loop(rmt_contex.hal.regs, channel, loop_en);
  339. RMT_EXIT_CRITICAL();
  340. return ESP_OK;
  341. }
  342. esp_err_t rmt_get_tx_loop_mode(rmt_channel_t channel, bool *loop_en)
  343. {
  344. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  345. RMT_ENTER_CRITICAL();
  346. *loop_en = rmt_ll_is_tx_loop_enabled(rmt_contex.hal.regs, channel);
  347. RMT_EXIT_CRITICAL();
  348. return ESP_OK;
  349. }
  350. esp_err_t rmt_set_rx_filter(rmt_channel_t channel, bool rx_filter_en, uint8_t thresh)
  351. {
  352. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  353. RMT_ENTER_CRITICAL();
  354. rmt_ll_rx_enable_filter(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), rx_filter_en);
  355. rmt_ll_rx_set_filter_thres(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), thresh);
  356. RMT_EXIT_CRITICAL();
  357. return ESP_OK;
  358. }
  359. esp_err_t rmt_set_source_clk(rmt_channel_t channel, rmt_source_clk_t base_clk)
  360. {
  361. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  362. RMT_CHECK(base_clk < RMT_BASECLK_MAX, RMT_BASECLK_ERROR_STR, ESP_ERR_INVALID_ARG);
  363. RMT_ENTER_CRITICAL();
  364. rmt_ll_set_counter_clock_src(rmt_contex.hal.regs, channel, base_clk, 0, 0, 0);
  365. RMT_EXIT_CRITICAL();
  366. return ESP_OK;
  367. }
  368. esp_err_t rmt_get_source_clk(rmt_channel_t channel, rmt_source_clk_t *src_clk)
  369. {
  370. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  371. RMT_ENTER_CRITICAL();
  372. *src_clk = (rmt_source_clk_t)rmt_ll_get_counter_clock_src(rmt_contex.hal.regs, channel);
  373. RMT_EXIT_CRITICAL();
  374. return ESP_OK;
  375. }
  376. esp_err_t rmt_set_idle_level(rmt_channel_t channel, bool idle_out_en, rmt_idle_level_t level)
  377. {
  378. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  379. RMT_CHECK(level < RMT_IDLE_LEVEL_MAX, "RMT IDLE LEVEL ERR", ESP_ERR_INVALID_ARG);
  380. RMT_ENTER_CRITICAL();
  381. rmt_ll_tx_enable_idle(rmt_contex.hal.regs, channel, idle_out_en);
  382. rmt_ll_tx_set_idle_level(rmt_contex.hal.regs, channel, level);
  383. RMT_EXIT_CRITICAL();
  384. return ESP_OK;
  385. }
  386. esp_err_t rmt_get_idle_level(rmt_channel_t channel, bool *idle_out_en, rmt_idle_level_t *level)
  387. {
  388. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  389. RMT_ENTER_CRITICAL();
  390. *idle_out_en = rmt_ll_is_tx_idle_enabled(rmt_contex.hal.regs, channel);
  391. *level = rmt_ll_tx_get_idle_level(rmt_contex.hal.regs, channel);
  392. RMT_EXIT_CRITICAL();
  393. return ESP_OK;
  394. }
  395. esp_err_t rmt_get_status(rmt_channel_t channel, uint32_t *status)
  396. {
  397. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  398. RMT_ENTER_CRITICAL();
  399. if (RMT_IS_RX_CHANNEL(channel)) {
  400. *status = rmt_ll_rx_get_channel_status(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  401. } else {
  402. *status = rmt_ll_tx_get_channel_status(rmt_contex.hal.regs, channel);
  403. }
  404. RMT_EXIT_CRITICAL();
  405. return ESP_OK;
  406. }
  407. void rmt_set_intr_enable_mask(uint32_t mask)
  408. {
  409. RMT_ENTER_CRITICAL();
  410. rmt_ll_set_intr_enable_mask(mask);
  411. RMT_EXIT_CRITICAL();
  412. }
  413. void rmt_clr_intr_enable_mask(uint32_t mask)
  414. {
  415. RMT_ENTER_CRITICAL();
  416. rmt_ll_clr_intr_enable_mask(mask);
  417. RMT_EXIT_CRITICAL();
  418. }
  419. esp_err_t rmt_set_rx_intr_en(rmt_channel_t channel, bool en)
  420. {
  421. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  422. RMT_ENTER_CRITICAL();
  423. rmt_ll_enable_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), en);
  424. RMT_EXIT_CRITICAL();
  425. return ESP_OK;
  426. }
  427. #if SOC_RMT_SUPPORT_RX_PINGPONG
  428. esp_err_t rmt_set_rx_thr_intr_en(rmt_channel_t channel, bool en, uint16_t evt_thresh)
  429. {
  430. RMT_CHECK(RMT_IS_RX_CHANNEL(channel) && channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  431. if (en) {
  432. uint32_t item_block_len = rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel)) * RMT_MEM_ITEM_NUM;
  433. RMT_CHECK(evt_thresh <= item_block_len, "RMT EVT THRESH ERR", ESP_ERR_INVALID_ARG);
  434. RMT_ENTER_CRITICAL();
  435. rmt_ll_rx_set_limit(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), evt_thresh);
  436. rmt_ll_enable_rx_thres_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), true);
  437. RMT_EXIT_CRITICAL();
  438. } else {
  439. RMT_ENTER_CRITICAL();
  440. rmt_ll_enable_rx_thres_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), false);
  441. RMT_EXIT_CRITICAL();
  442. }
  443. return ESP_OK;
  444. }
  445. #endif
  446. esp_err_t rmt_set_err_intr_en(rmt_channel_t channel, bool en)
  447. {
  448. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  449. RMT_ENTER_CRITICAL();
  450. if (RMT_IS_RX_CHANNEL(channel)) {
  451. rmt_ll_enable_rx_err_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), en);
  452. } else {
  453. rmt_ll_enable_tx_err_interrupt(rmt_contex.hal.regs, channel, en);
  454. }
  455. RMT_EXIT_CRITICAL();
  456. return ESP_OK;
  457. }
  458. esp_err_t rmt_set_tx_intr_en(rmt_channel_t channel, bool en)
  459. {
  460. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  461. RMT_ENTER_CRITICAL();
  462. rmt_ll_enable_tx_end_interrupt(rmt_contex.hal.regs, channel, en);
  463. RMT_EXIT_CRITICAL();
  464. return ESP_OK;
  465. }
  466. esp_err_t rmt_set_tx_thr_intr_en(rmt_channel_t channel, bool en, uint16_t evt_thresh)
  467. {
  468. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  469. if (en) {
  470. uint32_t item_block_len = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel) * RMT_MEM_ITEM_NUM;
  471. RMT_CHECK(evt_thresh <= item_block_len, "RMT EVT THRESH ERR", ESP_ERR_INVALID_ARG);
  472. RMT_ENTER_CRITICAL();
  473. rmt_ll_tx_set_limit(rmt_contex.hal.regs, channel, evt_thresh);
  474. rmt_ll_enable_tx_thres_interrupt(rmt_contex.hal.regs, channel, true);
  475. RMT_EXIT_CRITICAL();
  476. } else {
  477. RMT_ENTER_CRITICAL();
  478. rmt_ll_enable_tx_thres_interrupt(rmt_contex.hal.regs, channel, false);
  479. RMT_EXIT_CRITICAL();
  480. }
  481. return ESP_OK;
  482. }
  483. esp_err_t rmt_set_pin(rmt_channel_t channel, rmt_mode_t mode, gpio_num_t gpio_num)
  484. {
  485. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  486. RMT_CHECK(mode < RMT_MODE_MAX, RMT_MODE_ERROR_STR, ESP_ERR_INVALID_ARG);
  487. RMT_CHECK(((GPIO_IS_VALID_GPIO(gpio_num) && (mode == RMT_MODE_RX)) ||
  488. (GPIO_IS_VALID_OUTPUT_GPIO(gpio_num) && (mode == RMT_MODE_TX))),
  489. RMT_GPIO_ERROR_STR, ESP_ERR_INVALID_ARG);
  490. gpio_hal_iomux_func_sel(GPIO_PIN_MUX_REG[gpio_num], PIN_FUNC_GPIO);
  491. if (mode == RMT_MODE_TX) {
  492. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  493. gpio_set_direction(gpio_num, GPIO_MODE_OUTPUT);
  494. esp_rom_gpio_connect_out_signal(gpio_num, rmt_periph_signals.channels[channel].tx_sig, 0, 0);
  495. } else {
  496. RMT_CHECK(RMT_IS_RX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  497. gpio_set_direction(gpio_num, GPIO_MODE_INPUT);
  498. esp_rom_gpio_connect_in_signal(gpio_num, rmt_periph_signals.channels[channel].rx_sig, 0);
  499. }
  500. return ESP_OK;
  501. }
  502. static bool rmt_is_channel_number_valid(rmt_channel_t channel, uint8_t mode)
  503. {
  504. // RX mode
  505. if (mode == RMT_MODE_RX) {
  506. return RMT_IS_RX_CHANNEL(channel) && (channel < RMT_CHANNEL_MAX);
  507. }
  508. // TX mode
  509. return (channel >= 0) && RMT_IS_TX_CHANNEL(channel);
  510. }
  511. static esp_err_t rmt_internal_config(rmt_dev_t *dev, const rmt_config_t *rmt_param)
  512. {
  513. uint8_t mode = rmt_param->rmt_mode;
  514. uint8_t channel = rmt_param->channel;
  515. uint8_t gpio_num = rmt_param->gpio_num;
  516. uint8_t mem_cnt = rmt_param->mem_block_num;
  517. uint8_t clk_div = rmt_param->clk_div;
  518. uint32_t carrier_freq_hz = rmt_param->tx_config.carrier_freq_hz;
  519. bool carrier_en = rmt_param->tx_config.carrier_en;
  520. uint32_t rmt_source_clk_hz;
  521. RMT_CHECK(rmt_is_channel_number_valid(channel, mode), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  522. RMT_CHECK((mem_cnt + channel <= 8 && mem_cnt > 0), RMT_MEM_CNT_ERROR_STR, ESP_ERR_INVALID_ARG);
  523. RMT_CHECK((clk_div > 0), RMT_CLK_DIV_ERROR_STR, ESP_ERR_INVALID_ARG);
  524. if (mode == RMT_MODE_TX) {
  525. RMT_CHECK((!carrier_en || carrier_freq_hz > 0), "RMT carrier frequency can't be zero", ESP_ERR_INVALID_ARG);
  526. }
  527. RMT_ENTER_CRITICAL();
  528. rmt_ll_enable_mem_access(dev, true);
  529. if (rmt_param->flags & RMT_CHANNEL_FLAGS_AWARE_DFS) {
  530. #if SOC_RMT_SUPPORT_XTAL
  531. // clock src: XTAL_CLK
  532. rmt_source_clk_hz = rtc_clk_xtal_freq_get() * 1000000;
  533. rmt_ll_set_counter_clock_src(dev, channel, RMT_BASECLK_XTAL, 0, 0, 0);
  534. #elif SOC_RMT_SUPPORT_REF_TICK
  535. // clock src: REF_CLK
  536. rmt_source_clk_hz = REF_CLK_FREQ;
  537. rmt_ll_set_counter_clock_src(dev, channel, RMT_BASECLK_REF, 0, 0, 0);
  538. #endif
  539. } else {
  540. // clock src: APB_CLK
  541. rmt_source_clk_hz = APB_CLK_FREQ;
  542. rmt_ll_set_counter_clock_src(dev, channel, RMT_BASECLK_APB, 0, 0, 0);
  543. }
  544. RMT_EXIT_CRITICAL();
  545. #if SOC_RMT_SOURCE_CLK_INDEPENDENT
  546. s_rmt_source_clock_hz[channel] = rmt_source_clk_hz;
  547. #else
  548. if (s_rmt_source_clock_hz && rmt_source_clk_hz != s_rmt_source_clock_hz) {
  549. ESP_LOGW(RMT_TAG, "RMT clock source has been configured to %d by other channel, now reconfigure it to %d", s_rmt_source_clock_hz, rmt_source_clk_hz);
  550. }
  551. s_rmt_source_clock_hz = rmt_source_clk_hz;
  552. #endif
  553. ESP_LOGD(RMT_TAG, "rmt_source_clk_hz: %d\n", rmt_source_clk_hz);
  554. if (mode == RMT_MODE_TX) {
  555. uint16_t carrier_duty_percent = rmt_param->tx_config.carrier_duty_percent;
  556. uint8_t carrier_level = rmt_param->tx_config.carrier_level;
  557. uint8_t idle_level = rmt_param->tx_config.idle_level;
  558. RMT_ENTER_CRITICAL();
  559. rmt_ll_tx_set_counter_clock_div(dev, channel, clk_div);
  560. rmt_ll_tx_set_mem_blocks(dev, channel, mem_cnt);
  561. rmt_ll_tx_reset_pointer(dev, channel);
  562. rmt_ll_tx_enable_loop(dev, channel, rmt_param->tx_config.loop_en);
  563. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  564. if (rmt_param->tx_config.loop_en) {
  565. rmt_ll_tx_set_loop_count(dev, channel, rmt_param->tx_config.loop_count);
  566. }
  567. #endif
  568. /* always enable tx ping-pong */
  569. rmt_ll_tx_enable_pingpong(dev, channel, true);
  570. /*Set idle level */
  571. rmt_ll_tx_enable_idle(dev, channel, rmt_param->tx_config.idle_output_en);
  572. rmt_ll_tx_set_idle_level(dev, channel, idle_level);
  573. /*Set carrier*/
  574. rmt_ll_tx_enable_carrier_modulation(dev, channel, carrier_en);
  575. if (carrier_en) {
  576. uint32_t duty_div, duty_h, duty_l;
  577. duty_div = rmt_source_clk_hz / carrier_freq_hz;
  578. duty_h = duty_div * carrier_duty_percent / 100;
  579. duty_l = duty_div - duty_h;
  580. rmt_ll_tx_set_carrier_level(dev, channel, carrier_level);
  581. rmt_ll_tx_set_carrier_high_low_ticks(dev, channel, duty_h, duty_l);
  582. } else {
  583. rmt_ll_tx_set_carrier_level(dev, channel, 0);
  584. rmt_ll_tx_set_carrier_high_low_ticks(dev, channel, 0, 0);
  585. }
  586. RMT_EXIT_CRITICAL();
  587. ESP_LOGD(RMT_TAG, "Rmt Tx Channel %u|Gpio %u|Sclk_Hz %u|Div %u|Carrier_Hz %u|Duty %u",
  588. channel, gpio_num, rmt_source_clk_hz, clk_div, carrier_freq_hz, carrier_duty_percent);
  589. } else if (RMT_MODE_RX == mode) {
  590. uint8_t filter_cnt = rmt_param->rx_config.filter_ticks_thresh;
  591. uint16_t threshold = rmt_param->rx_config.idle_threshold;
  592. RMT_ENTER_CRITICAL();
  593. rmt_ll_rx_set_counter_clock_div(dev, RMT_DECODE_RX_CHANNEL(channel), clk_div);
  594. rmt_ll_rx_set_mem_blocks(dev, RMT_DECODE_RX_CHANNEL(channel), mem_cnt);
  595. rmt_ll_rx_reset_pointer(dev, RMT_DECODE_RX_CHANNEL(channel));
  596. rmt_ll_rx_set_mem_owner(dev, RMT_DECODE_RX_CHANNEL(channel), RMT_MEM_OWNER_HW);
  597. /*Set idle threshold*/
  598. rmt_ll_rx_set_idle_thres(dev, RMT_DECODE_RX_CHANNEL(channel), threshold);
  599. /* Set RX filter */
  600. rmt_ll_rx_set_filter_thres(dev, RMT_DECODE_RX_CHANNEL(channel), filter_cnt);
  601. rmt_ll_rx_enable_filter(dev, RMT_DECODE_RX_CHANNEL(channel), rmt_param->rx_config.filter_en);
  602. #if SOC_RMT_SUPPORT_RX_PINGPONG
  603. /* always enable rx ping-pong */
  604. rmt_ll_rx_enable_pingpong(dev, RMT_DECODE_RX_CHANNEL(channel), true);
  605. #endif
  606. #if SOC_RMT_SUPPORT_RX_DEMODULATION
  607. rmt_ll_rx_enable_carrier_demodulation(dev, RMT_DECODE_RX_CHANNEL(channel), rmt_param->rx_config.rm_carrier);
  608. if (rmt_param->rx_config.rm_carrier) {
  609. uint32_t duty_total = rmt_source_clk_hz / rmt_ll_rx_get_counter_clock_div(dev, RMT_DECODE_RX_CHANNEL(channel)) / rmt_param->rx_config.carrier_freq_hz;
  610. uint32_t duty_high = duty_total * rmt_param->rx_config.carrier_duty_percent / 100;
  611. // there could be residual in timing the carrier pulse, so double enlarge the theoretical value
  612. rmt_ll_rx_set_carrier_high_low_ticks(dev, RMT_DECODE_RX_CHANNEL(channel), duty_high * 2, (duty_total - duty_high) * 2);
  613. rmt_ll_rx_set_carrier_level(dev, RMT_DECODE_RX_CHANNEL(channel), rmt_param->rx_config.carrier_level);
  614. }
  615. #endif
  616. RMT_EXIT_CRITICAL();
  617. ESP_LOGD(RMT_TAG, "Rmt Rx Channel %u|Gpio %u|Sclk_Hz %u|Div %u|Thresold %u|Filter %u",
  618. channel, gpio_num, rmt_source_clk_hz, clk_div, threshold, filter_cnt);
  619. }
  620. return ESP_OK;
  621. }
  622. esp_err_t rmt_config(const rmt_config_t *rmt_param)
  623. {
  624. rmt_module_enable();
  625. RMT_CHECK(rmt_set_pin(rmt_param->channel, rmt_param->rmt_mode, rmt_param->gpio_num) == ESP_OK,
  626. "set gpio for RMT driver failed", ESP_ERR_INVALID_ARG);
  627. RMT_CHECK(rmt_internal_config(&RMT, rmt_param) == ESP_OK,
  628. "initialize RMT driver failed", ESP_ERR_INVALID_ARG);
  629. return ESP_OK;
  630. }
  631. static void IRAM_ATTR rmt_fill_memory(rmt_channel_t channel, const rmt_item32_t *item,
  632. uint16_t item_num, uint16_t mem_offset)
  633. {
  634. RMT_ENTER_CRITICAL();
  635. rmt_ll_write_memory(rmt_contex.hal.mem, channel, item, item_num, mem_offset);
  636. RMT_EXIT_CRITICAL();
  637. }
  638. esp_err_t rmt_fill_tx_items(rmt_channel_t channel, const rmt_item32_t *item, uint16_t item_num, uint16_t mem_offset)
  639. {
  640. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, (0));
  641. RMT_CHECK((item != NULL), RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  642. RMT_CHECK((item_num > 0), RMT_DRIVER_LENGTH_ERROR_STR, ESP_ERR_INVALID_ARG);
  643. /*Each block has 64 x 32 bits of data*/
  644. uint8_t mem_cnt = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel);
  645. RMT_CHECK((mem_cnt * RMT_MEM_ITEM_NUM >= item_num), RMT_WR_MEM_OVF_ERROR_STR, ESP_ERR_INVALID_ARG);
  646. rmt_fill_memory(channel, item, item_num, mem_offset);
  647. return ESP_OK;
  648. }
  649. esp_err_t rmt_isr_register(void (*fn)(void *), void *arg, int intr_alloc_flags, rmt_isr_handle_t *handle)
  650. {
  651. RMT_CHECK((fn != NULL), RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  652. RMT_CHECK(rmt_contex.rmt_driver_channels == 0, "RMT driver installed, can not install generic ISR handler", ESP_FAIL);
  653. return esp_intr_alloc(rmt_periph_signals.irq, intr_alloc_flags, fn, arg, handle);
  654. }
  655. esp_err_t rmt_isr_deregister(rmt_isr_handle_t handle)
  656. {
  657. return esp_intr_free(handle);
  658. }
  659. static int IRAM_ATTR rmt_rx_get_mem_len_in_isr(rmt_channel_t channel)
  660. {
  661. int block_num = rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, channel);
  662. int item_block_len = block_num * RMT_MEM_ITEM_NUM;
  663. volatile rmt_item32_t *data = (rmt_item32_t *)RMTMEM.chan[RMT_ENCODE_RX_CHANNEL(channel)].data32;
  664. int idx;
  665. for (idx = 0; idx < item_block_len; idx++) {
  666. if (data[idx].duration0 == 0) {
  667. return idx;
  668. } else if (data[idx].duration1 == 0) {
  669. return idx + 1;
  670. }
  671. }
  672. return idx;
  673. }
  674. static void IRAM_ATTR rmt_driver_isr_default(void *arg)
  675. {
  676. uint32_t status = 0;
  677. rmt_item32_t volatile *addr = NULL;
  678. uint8_t channel = 0;
  679. rmt_hal_context_t *hal = (rmt_hal_context_t *)arg;
  680. portBASE_TYPE HPTaskAwoken = pdFALSE;
  681. // Tx end interrupt
  682. status = rmt_ll_get_tx_end_interrupt_status(hal->regs);
  683. while (status) {
  684. channel = __builtin_ffs(status) - 1;
  685. status &= ~(1 << channel);
  686. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  687. if (p_rmt) {
  688. xSemaphoreGiveFromISR(p_rmt->tx_sem, &HPTaskAwoken);
  689. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  690. p_rmt->tx_data = NULL;
  691. p_rmt->tx_len_rem = 0;
  692. p_rmt->tx_offset = 0;
  693. p_rmt->tx_sub_len = 0;
  694. p_rmt->sample_cur = NULL;
  695. p_rmt->translator = false;
  696. if (rmt_contex.rmt_tx_end_callback.function != NULL) {
  697. rmt_contex.rmt_tx_end_callback.function(channel, rmt_contex.rmt_tx_end_callback.arg);
  698. }
  699. }
  700. rmt_ll_clear_tx_end_interrupt(hal->regs, channel);
  701. }
  702. // Tx thres interrupt
  703. status = rmt_ll_get_tx_thres_interrupt_status(hal->regs);
  704. while (status) {
  705. channel = __builtin_ffs(status) - 1;
  706. status &= ~(1 << channel);
  707. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  708. if (p_rmt) {
  709. if (p_rmt->translator) {
  710. if (p_rmt->sample_size_remain > 0) {
  711. size_t translated_size = 0;
  712. p_rmt->sample_to_rmt((void *)p_rmt->sample_cur,
  713. p_rmt->tx_buf,
  714. p_rmt->sample_size_remain,
  715. p_rmt->tx_sub_len,
  716. &translated_size,
  717. &p_rmt->tx_len_rem);
  718. p_rmt->sample_size_remain -= translated_size;
  719. p_rmt->sample_cur += translated_size;
  720. p_rmt->tx_data = p_rmt->tx_buf;
  721. } else {
  722. p_rmt->sample_cur = NULL;
  723. p_rmt->translator = false;
  724. }
  725. }
  726. const rmt_item32_t *pdata = p_rmt->tx_data;
  727. size_t len_rem = p_rmt->tx_len_rem;
  728. if (len_rem >= p_rmt->tx_sub_len) {
  729. rmt_fill_memory(channel, pdata, p_rmt->tx_sub_len, p_rmt->tx_offset);
  730. p_rmt->tx_data += p_rmt->tx_sub_len;
  731. p_rmt->tx_len_rem -= p_rmt->tx_sub_len;
  732. } else if (len_rem == 0) {
  733. rmt_item32_t stop_data = {0};
  734. rmt_ll_write_memory(rmt_contex.hal.mem, channel, &stop_data, 1, p_rmt->tx_offset);
  735. } else {
  736. rmt_fill_memory(channel, pdata, len_rem, p_rmt->tx_offset);
  737. rmt_item32_t stop_data = {0};
  738. rmt_ll_write_memory(rmt_contex.hal.mem, channel, &stop_data, 1, p_rmt->tx_offset + len_rem);
  739. p_rmt->tx_data += len_rem;
  740. p_rmt->tx_len_rem -= len_rem;
  741. }
  742. if (p_rmt->tx_offset == 0) {
  743. p_rmt->tx_offset = p_rmt->tx_sub_len;
  744. } else {
  745. p_rmt->tx_offset = 0;
  746. }
  747. }
  748. rmt_ll_clear_tx_thres_interrupt(hal->regs, channel);
  749. }
  750. // Rx end interrupt
  751. status = rmt_ll_get_rx_end_interrupt_status(hal->regs);
  752. while (status) {
  753. channel = __builtin_ffs(status) - 1;
  754. status &= ~(1 << channel);
  755. rmt_obj_t *p_rmt = p_rmt_obj[RMT_ENCODE_RX_CHANNEL(channel)];
  756. if (p_rmt) {
  757. rmt_ll_rx_enable(rmt_contex.hal.regs, channel, false);
  758. int item_len = rmt_rx_get_mem_len_in_isr(channel);
  759. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, channel, RMT_MEM_OWNER_SW);
  760. if (p_rmt->rx_buf) {
  761. addr = RMTMEM.chan[RMT_ENCODE_RX_CHANNEL(channel)].data32;
  762. #if SOC_RMT_SUPPORT_RX_PINGPONG
  763. if (item_len > p_rmt->rx_item_start_idx) {
  764. item_len = item_len - p_rmt->rx_item_start_idx;
  765. }
  766. memcpy((void *)(p_rmt->rx_item_buf + p_rmt->rx_item_len), (void *)(addr + p_rmt->rx_item_start_idx), item_len * 4);
  767. p_rmt->rx_item_len += item_len;
  768. BaseType_t res = xRingbufferSendFromISR(p_rmt->rx_buf, (void *)(p_rmt->rx_item_buf), p_rmt->rx_item_len * 4, &HPTaskAwoken);
  769. #else
  770. BaseType_t res = xRingbufferSendFromISR(p_rmt->rx_buf, (void *)addr, item_len * 4, &HPTaskAwoken);
  771. #endif
  772. if (res == pdFALSE) {
  773. ESP_EARLY_LOGE(RMT_TAG, "RMT RX BUFFER FULL");
  774. }
  775. } else {
  776. ESP_EARLY_LOGE(RMT_TAG, "RMT RX BUFFER ERROR");
  777. }
  778. #if SOC_RMT_SUPPORT_RX_PINGPONG
  779. p_rmt->rx_item_start_idx = 0;
  780. p_rmt->rx_item_len = 0;
  781. memset((void *)p_rmt->rx_item_buf, 0, p_rmt->rx_item_buf_size);
  782. #endif
  783. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, channel);
  784. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, channel, RMT_MEM_OWNER_HW);
  785. rmt_ll_rx_enable(rmt_contex.hal.regs, channel, true);
  786. }
  787. rmt_ll_clear_rx_end_interrupt(hal->regs, channel);
  788. }
  789. #if SOC_RMT_SUPPORT_RX_PINGPONG
  790. // Rx thres interrupt
  791. status = rmt_ll_get_rx_thres_interrupt_status(hal->regs);
  792. while (status) {
  793. channel = __builtin_ffs(status) - 1;
  794. status &= ~(1 << channel);
  795. rmt_obj_t *p_rmt = p_rmt_obj[RMT_ENCODE_RX_CHANNEL(channel)];
  796. int mem_item_size = rmt_ll_rx_get_mem_blocks(rmt_contex.hal.regs, channel) * RMT_MEM_ITEM_NUM;
  797. int rx_thres_lim = rmt_ll_rx_get_limit(rmt_contex.hal.regs, channel);
  798. int item_len = (p_rmt->rx_item_start_idx == 0) ? rx_thres_lim : (mem_item_size - rx_thres_lim);
  799. if ((p_rmt->rx_item_len + item_len) < (p_rmt->rx_item_buf_size / 4)) {
  800. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, channel, RMT_MEM_OWNER_SW);
  801. memcpy((void *)(p_rmt->rx_item_buf + p_rmt->rx_item_len), (void *)(RMTMEM.chan[RMT_ENCODE_RX_CHANNEL(channel)].data32 + p_rmt->rx_item_start_idx), item_len * 4);
  802. rmt_ll_rx_set_mem_owner(rmt_contex.hal.regs, channel, RMT_MEM_OWNER_HW);
  803. p_rmt->rx_item_len += item_len;
  804. p_rmt->rx_item_start_idx += item_len;
  805. if (p_rmt->rx_item_start_idx >= mem_item_size) {
  806. p_rmt->rx_item_start_idx = 0;
  807. }
  808. } else {
  809. ESP_EARLY_LOGE(RMT_TAG, "---RX buffer too small: %d", sizeof(p_rmt->rx_item_buf));
  810. }
  811. rmt_ll_clear_rx_thres_interrupt(hal->regs, channel);
  812. }
  813. #endif
  814. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  815. // loop count interrupt
  816. status = rmt_ll_get_tx_loop_interrupt_status(hal->regs);
  817. while (status) {
  818. channel = __builtin_ffs(status) - 1;
  819. status &= ~(1 << channel);
  820. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  821. if (p_rmt) {
  822. xSemaphoreGiveFromISR(p_rmt->tx_sem, &HPTaskAwoken);
  823. if (rmt_contex.rmt_tx_end_callback.function != NULL) {
  824. rmt_contex.rmt_tx_end_callback.function(channel, rmt_contex.rmt_tx_end_callback.arg);
  825. }
  826. }
  827. rmt_ll_clear_tx_loop_interrupt(hal->regs, channel);
  828. }
  829. #endif
  830. // RX Err interrupt
  831. status = rmt_ll_get_rx_err_interrupt_status(hal->regs);
  832. while (status) {
  833. channel = __builtin_ffs(status) - 1;
  834. status &= ~(1 << channel);
  835. rmt_obj_t *p_rmt = p_rmt_obj[RMT_ENCODE_RX_CHANNEL(channel)];
  836. if (p_rmt) {
  837. // Reset the receiver's write/read addresses to prevent endless err interrupts.
  838. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, channel);
  839. ESP_EARLY_LOGD(RMT_TAG, "RMT RX channel %d error", channel);
  840. ESP_EARLY_LOGD(RMT_TAG, "status: 0x%08x", rmt_ll_rx_get_channel_status(rmt_contex.hal.regs, channel));
  841. }
  842. rmt_ll_clear_rx_err_interrupt(hal->regs, channel);
  843. }
  844. // TX Err interrupt
  845. status = rmt_ll_get_tx_err_interrupt_status(hal->regs);
  846. while (status) {
  847. channel = __builtin_ffs(status) - 1;
  848. status &= ~(1 << channel);
  849. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  850. if (p_rmt) {
  851. // Reset the transmitter's write/read addresses to prevent endless err interrupts.
  852. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  853. ESP_EARLY_LOGD(RMT_TAG, "RMT TX channel %d error", channel);
  854. ESP_EARLY_LOGD(RMT_TAG, "status: 0x%08x", rmt_ll_tx_get_channel_status(rmt_contex.hal.regs, channel));
  855. }
  856. rmt_ll_clear_tx_err_interrupt(hal->regs, channel);
  857. }
  858. if (HPTaskAwoken == pdTRUE) {
  859. portYIELD_FROM_ISR();
  860. }
  861. }
  862. esp_err_t rmt_driver_uninstall(rmt_channel_t channel)
  863. {
  864. esp_err_t err = ESP_OK;
  865. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  866. RMT_CHECK((rmt_contex.rmt_driver_channels & BIT(channel)) != 0, "No RMT driver for this channel", ESP_ERR_INVALID_STATE);
  867. if (p_rmt_obj[channel] == NULL) {
  868. return ESP_OK;
  869. }
  870. //Avoid blocking here(when the interrupt is disabled and do not wait tx done).
  871. if (p_rmt_obj[channel]->wait_done) {
  872. xSemaphoreTake(p_rmt_obj[channel]->tx_sem, portMAX_DELAY);
  873. }
  874. RMT_ENTER_CRITICAL();
  875. // check channel's working mode
  876. if (p_rmt_obj[channel]->rx_buf) {
  877. rmt_ll_enable_rx_end_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), 0);
  878. rmt_ll_enable_rx_err_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), 0);
  879. #if SOC_RMT_SUPPORT_RX_PINGPONG
  880. rmt_ll_enable_rx_thres_interrupt(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel), 0);
  881. #endif
  882. } else {
  883. rmt_ll_enable_tx_end_interrupt(rmt_contex.hal.regs, channel, 0);
  884. rmt_ll_enable_tx_err_interrupt(rmt_contex.hal.regs, channel, 0);
  885. rmt_ll_enable_tx_thres_interrupt(rmt_contex.hal.regs, channel, false);
  886. }
  887. RMT_EXIT_CRITICAL();
  888. _lock_acquire_recursive(&(rmt_contex.rmt_driver_isr_lock));
  889. rmt_contex.rmt_driver_channels &= ~BIT(channel);
  890. if (rmt_contex.rmt_driver_channels == 0) {
  891. rmt_module_disable();
  892. // all channels have driver disabled
  893. err = rmt_isr_deregister(rmt_contex.rmt_driver_intr_handle);
  894. rmt_contex.rmt_driver_intr_handle = NULL;
  895. }
  896. _lock_release_recursive(&(rmt_contex.rmt_driver_isr_lock));
  897. if (err != ESP_OK) {
  898. return err;
  899. }
  900. if (p_rmt_obj[channel]->tx_sem) {
  901. vSemaphoreDelete(p_rmt_obj[channel]->tx_sem);
  902. p_rmt_obj[channel]->tx_sem = NULL;
  903. }
  904. if (p_rmt_obj[channel]->rx_buf) {
  905. vRingbufferDelete(p_rmt_obj[channel]->rx_buf);
  906. p_rmt_obj[channel]->rx_buf = NULL;
  907. }
  908. if (p_rmt_obj[channel]->tx_buf) {
  909. free(p_rmt_obj[channel]->tx_buf);
  910. p_rmt_obj[channel]->tx_buf = NULL;
  911. }
  912. if (p_rmt_obj[channel]->sample_to_rmt) {
  913. p_rmt_obj[channel]->sample_to_rmt = NULL;
  914. }
  915. #if SOC_RMT_SUPPORT_RX_PINGPONG
  916. if (p_rmt_obj[channel]->rx_item_buf) {
  917. free(p_rmt_obj[channel]->rx_item_buf);
  918. p_rmt_obj[channel]->rx_item_buf = NULL;
  919. p_rmt_obj[channel]->rx_item_buf_size = 0;
  920. }
  921. #endif
  922. free(p_rmt_obj[channel]);
  923. p_rmt_obj[channel] = NULL;
  924. return ESP_OK;
  925. }
  926. esp_err_t rmt_driver_install(rmt_channel_t channel, size_t rx_buf_size, int intr_alloc_flags)
  927. {
  928. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  929. RMT_CHECK((rmt_contex.rmt_driver_channels & BIT(channel)) == 0,
  930. "RMT driver already installed for channel", ESP_ERR_INVALID_STATE);
  931. esp_err_t err = ESP_OK;
  932. if (p_rmt_obj[channel] != NULL) {
  933. ESP_LOGD(RMT_TAG, "RMT driver already installed");
  934. return ESP_ERR_INVALID_STATE;
  935. }
  936. #if !CONFIG_SPIRAM_USE_MALLOC
  937. p_rmt_obj[channel] = calloc(1, sizeof(rmt_obj_t));
  938. #else
  939. if (!(intr_alloc_flags & ESP_INTR_FLAG_IRAM)) {
  940. p_rmt_obj[channel] = calloc(1, sizeof(rmt_obj_t));
  941. } else {
  942. p_rmt_obj[channel] = heap_caps_calloc(1, sizeof(rmt_obj_t), MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT);
  943. }
  944. #endif
  945. if (p_rmt_obj[channel] == NULL) {
  946. ESP_LOGE(RMT_TAG, "RMT driver malloc error");
  947. return ESP_ERR_NO_MEM;
  948. }
  949. p_rmt_obj[channel]->tx_len_rem = 0;
  950. p_rmt_obj[channel]->tx_data = NULL;
  951. p_rmt_obj[channel]->channel = channel;
  952. p_rmt_obj[channel]->tx_offset = 0;
  953. p_rmt_obj[channel]->tx_sub_len = 0;
  954. p_rmt_obj[channel]->wait_done = false;
  955. p_rmt_obj[channel]->translator = false;
  956. p_rmt_obj[channel]->sample_to_rmt = NULL;
  957. if (p_rmt_obj[channel]->tx_sem == NULL) {
  958. #if !CONFIG_SPIRAM_USE_MALLOC
  959. p_rmt_obj[channel]->tx_sem = xSemaphoreCreateBinary();
  960. #else
  961. p_rmt_obj[channel]->intr_alloc_flags = intr_alloc_flags;
  962. if (!(intr_alloc_flags & ESP_INTR_FLAG_IRAM)) {
  963. p_rmt_obj[channel]->tx_sem = xSemaphoreCreateBinary();
  964. } else {
  965. p_rmt_obj[channel]->tx_sem = xSemaphoreCreateBinaryStatic(&p_rmt_obj[channel]->tx_sem_buffer);
  966. }
  967. #endif
  968. xSemaphoreGive(p_rmt_obj[channel]->tx_sem);
  969. }
  970. if (p_rmt_obj[channel]->rx_buf == NULL && rx_buf_size > 0) {
  971. p_rmt_obj[channel]->rx_buf = xRingbufferCreate(rx_buf_size, RINGBUF_TYPE_NOSPLIT);
  972. }
  973. #if SOC_RMT_SUPPORT_RX_PINGPONG
  974. if (p_rmt_obj[channel]->rx_item_buf == NULL && rx_buf_size > 0) {
  975. #if !CONFIG_SPIRAM_USE_MALLOC
  976. p_rmt_obj[channel]->rx_item_buf = calloc(1, rx_buf_size);
  977. #else
  978. if (!(p_rmt_obj[channel]->intr_alloc_flags & ESP_INTR_FLAG_IRAM)) {
  979. p_rmt_obj[channel]->rx_item_buf = calloc(1, rx_buf_size);
  980. } else {
  981. p_rmt_obj[channel]->rx_item_buf = heap_caps_calloc(1, rx_buf_size, MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT);
  982. }
  983. #endif
  984. if (p_rmt_obj[channel]->rx_item_buf == NULL) {
  985. ESP_LOGE(RMT_TAG, "RMT malloc fail");
  986. return ESP_FAIL;
  987. }
  988. p_rmt_obj[channel]->rx_item_buf_size = rx_buf_size;
  989. }
  990. #endif
  991. _lock_acquire_recursive(&(rmt_contex.rmt_driver_isr_lock));
  992. if (rmt_contex.rmt_driver_channels == 0) {
  993. // first RMT channel using driver
  994. err = rmt_isr_register(rmt_driver_isr_default, &rmt_contex.hal, intr_alloc_flags, &(rmt_contex.rmt_driver_intr_handle));
  995. }
  996. if (err == ESP_OK) {
  997. rmt_contex.rmt_driver_channels |= BIT(channel);
  998. }
  999. _lock_release_recursive(&(rmt_contex.rmt_driver_isr_lock));
  1000. rmt_module_enable();
  1001. if (RMT_IS_RX_CHANNEL(channel)) {
  1002. rmt_hal_rx_channel_reset(&rmt_contex.hal, RMT_DECODE_RX_CHANNEL(channel));
  1003. } else {
  1004. rmt_hal_tx_channel_reset(&rmt_contex.hal, channel);
  1005. }
  1006. return err;
  1007. }
  1008. esp_err_t rmt_write_items(rmt_channel_t channel, const rmt_item32_t *rmt_item, int item_num, bool wait_tx_done)
  1009. {
  1010. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1011. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1012. RMT_CHECK(rmt_item != NULL, RMT_ADDR_ERROR_STR, ESP_FAIL);
  1013. RMT_CHECK(item_num > 0, RMT_DRIVER_LENGTH_ERROR_STR, ESP_ERR_INVALID_ARG);
  1014. #if CONFIG_SPIRAM_USE_MALLOC
  1015. if (p_rmt_obj[channel]->intr_alloc_flags & ESP_INTR_FLAG_IRAM) {
  1016. if (!esp_ptr_internal(rmt_item)) {
  1017. ESP_LOGE(RMT_TAG, RMT_PSRAM_BUFFER_WARN_STR);
  1018. return ESP_ERR_INVALID_ARG;
  1019. }
  1020. }
  1021. #endif
  1022. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  1023. int block_num = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel);
  1024. int item_block_len = block_num * RMT_MEM_ITEM_NUM;
  1025. int item_sub_len = block_num * RMT_MEM_ITEM_NUM / 2;
  1026. int len_rem = item_num;
  1027. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1028. // fill the memory block first
  1029. if (item_num >= item_block_len) {
  1030. rmt_fill_memory(channel, rmt_item, item_block_len, 0);
  1031. len_rem -= item_block_len;
  1032. rmt_set_tx_loop_mode(channel, false);
  1033. rmt_set_tx_thr_intr_en(channel, 1, item_sub_len);
  1034. p_rmt->tx_data = rmt_item + item_block_len;
  1035. p_rmt->tx_len_rem = len_rem;
  1036. p_rmt->tx_offset = 0;
  1037. p_rmt->tx_sub_len = item_sub_len;
  1038. } else {
  1039. rmt_fill_memory(channel, rmt_item, len_rem, 0);
  1040. rmt_item32_t stop_data = {0};
  1041. rmt_ll_write_memory(rmt_contex.hal.mem, channel, &stop_data, 1, len_rem);
  1042. p_rmt->tx_len_rem = 0;
  1043. }
  1044. rmt_tx_start(channel, true);
  1045. p_rmt->wait_done = wait_tx_done;
  1046. if (wait_tx_done) {
  1047. // wait loop done
  1048. if (rmt_ll_is_tx_loop_enabled(rmt_contex.hal.regs, channel)) {
  1049. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  1050. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1051. xSemaphoreGive(p_rmt->tx_sem);
  1052. #endif
  1053. } else {
  1054. // wait tx end
  1055. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1056. xSemaphoreGive(p_rmt->tx_sem);
  1057. }
  1058. }
  1059. return ESP_OK;
  1060. }
  1061. esp_err_t rmt_wait_tx_done(rmt_channel_t channel, TickType_t wait_time)
  1062. {
  1063. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1064. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1065. if (xSemaphoreTake(p_rmt_obj[channel]->tx_sem, wait_time) == pdTRUE) {
  1066. p_rmt_obj[channel]->wait_done = false;
  1067. xSemaphoreGive(p_rmt_obj[channel]->tx_sem);
  1068. return ESP_OK;
  1069. } else {
  1070. if (wait_time != 0) {
  1071. // Don't emit error message if just polling.
  1072. ESP_LOGE(RMT_TAG, "Timeout on wait_tx_done");
  1073. }
  1074. return ESP_ERR_TIMEOUT;
  1075. }
  1076. }
  1077. esp_err_t rmt_get_ringbuf_handle(rmt_channel_t channel, RingbufHandle_t *buf_handle)
  1078. {
  1079. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1080. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1081. RMT_CHECK(buf_handle != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  1082. *buf_handle = p_rmt_obj[channel]->rx_buf;
  1083. return ESP_OK;
  1084. }
  1085. rmt_tx_end_callback_t rmt_register_tx_end_callback(rmt_tx_end_fn_t function, void *arg)
  1086. {
  1087. rmt_tx_end_callback_t previous = rmt_contex.rmt_tx_end_callback;
  1088. rmt_contex.rmt_tx_end_callback.function = function;
  1089. rmt_contex.rmt_tx_end_callback.arg = arg;
  1090. return previous;
  1091. }
  1092. esp_err_t rmt_translator_init(rmt_channel_t channel, sample_to_rmt_t fn)
  1093. {
  1094. RMT_CHECK(fn != NULL, RMT_TRANSLATOR_NULL_STR, ESP_ERR_INVALID_ARG);
  1095. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1096. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1097. const uint32_t block_size = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel) *
  1098. RMT_MEM_ITEM_NUM * sizeof(rmt_item32_t);
  1099. if (p_rmt_obj[channel]->tx_buf == NULL) {
  1100. #if !CONFIG_SPIRAM_USE_MALLOC
  1101. p_rmt_obj[channel]->tx_buf = (rmt_item32_t *)malloc(block_size);
  1102. #else
  1103. if (p_rmt_obj[channel]->intr_alloc_flags & ESP_INTR_FLAG_IRAM) {
  1104. p_rmt_obj[channel]->tx_buf = (rmt_item32_t *)malloc(block_size);
  1105. } else {
  1106. p_rmt_obj[channel]->tx_buf = (rmt_item32_t *)heap_caps_calloc(1, block_size, MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT);
  1107. }
  1108. #endif
  1109. if (p_rmt_obj[channel]->tx_buf == NULL) {
  1110. ESP_LOGE(RMT_TAG, "RMT translator buffer create fail");
  1111. return ESP_FAIL;
  1112. }
  1113. }
  1114. p_rmt_obj[channel]->sample_to_rmt = fn;
  1115. p_rmt_obj[channel]->tx_context = NULL;
  1116. p_rmt_obj[channel]->sample_size_remain = 0;
  1117. p_rmt_obj[channel]->sample_cur = NULL;
  1118. ESP_LOGD(RMT_TAG, "RMT translator init done");
  1119. return ESP_OK;
  1120. }
  1121. esp_err_t rmt_translator_set_context(rmt_channel_t channel, void *context)
  1122. {
  1123. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1124. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1125. p_rmt_obj[channel]->tx_context = context;
  1126. return ESP_OK;
  1127. }
  1128. esp_err_t rmt_translator_get_context(const size_t *item_num, void **context)
  1129. {
  1130. RMT_CHECK(item_num && context, "invalid arguments", ESP_ERR_INVALID_ARG);
  1131. // the address of tx_len_rem is directlly passed to the callback,
  1132. // so it's possible to get the object address from that
  1133. rmt_obj_t *obj = __containerof(item_num, rmt_obj_t, tx_len_rem);
  1134. *context = obj->tx_context;
  1135. return ESP_OK;
  1136. }
  1137. esp_err_t rmt_write_sample(rmt_channel_t channel, const uint8_t *src, size_t src_size, bool wait_tx_done)
  1138. {
  1139. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1140. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  1141. RMT_CHECK(p_rmt_obj[channel]->sample_to_rmt != NULL, RMT_TRANSLATOR_UNINIT_STR, ESP_FAIL);
  1142. #if CONFIG_SPIRAM_USE_MALLOC
  1143. if (p_rmt_obj[channel]->intr_alloc_flags & ESP_INTR_FLAG_IRAM) {
  1144. if (!esp_ptr_internal(src)) {
  1145. ESP_LOGE(RMT_TAG, RMT_PSRAM_BUFFER_WARN_STR);
  1146. return ESP_ERR_INVALID_ARG;
  1147. }
  1148. }
  1149. #endif
  1150. size_t translated_size = 0;
  1151. rmt_obj_t *p_rmt = p_rmt_obj[channel];
  1152. const uint32_t item_block_len = rmt_ll_tx_get_mem_blocks(rmt_contex.hal.regs, channel) * RMT_MEM_ITEM_NUM;
  1153. const uint32_t item_sub_len = item_block_len / 2;
  1154. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1155. p_rmt->sample_to_rmt((void *)src, p_rmt->tx_buf, src_size, item_block_len, &translated_size, &p_rmt->tx_len_rem);
  1156. p_rmt->sample_size_remain = src_size - translated_size;
  1157. p_rmt->sample_cur = src + translated_size;
  1158. rmt_fill_memory(channel, p_rmt->tx_buf, p_rmt->tx_len_rem, 0);
  1159. if (p_rmt->tx_len_rem == item_block_len) {
  1160. rmt_set_tx_thr_intr_en(channel, 1, item_sub_len);
  1161. p_rmt->tx_data = p_rmt->tx_buf;
  1162. p_rmt->tx_offset = 0;
  1163. p_rmt->tx_sub_len = item_sub_len;
  1164. p_rmt->translator = true;
  1165. } else {
  1166. rmt_item32_t stop_data = {0};
  1167. rmt_ll_write_memory(rmt_contex.hal.mem, channel, &stop_data, 1, p_rmt->tx_len_rem);
  1168. p_rmt->tx_len_rem = 0;
  1169. p_rmt->sample_cur = NULL;
  1170. p_rmt->translator = false;
  1171. }
  1172. rmt_tx_start(channel, true);
  1173. p_rmt->wait_done = wait_tx_done;
  1174. if (wait_tx_done) {
  1175. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  1176. xSemaphoreGive(p_rmt->tx_sem);
  1177. }
  1178. return ESP_OK;
  1179. }
  1180. esp_err_t rmt_get_channel_status(rmt_channel_status_result_t *channel_status)
  1181. {
  1182. RMT_CHECK(channel_status != NULL, RMT_PARAM_ERR_STR, ESP_ERR_INVALID_ARG);
  1183. for (int i = 0; i < RMT_CHANNEL_MAX; i++) {
  1184. channel_status->status[i] = RMT_CHANNEL_UNINIT;
  1185. if (p_rmt_obj[i] != NULL) {
  1186. if (p_rmt_obj[i]->tx_sem != NULL) {
  1187. if (xSemaphoreTake(p_rmt_obj[i]->tx_sem, (TickType_t)0) == pdTRUE) {
  1188. channel_status->status[i] = RMT_CHANNEL_IDLE;
  1189. xSemaphoreGive(p_rmt_obj[i]->tx_sem);
  1190. } else {
  1191. channel_status->status[i] = RMT_CHANNEL_BUSY;
  1192. }
  1193. }
  1194. }
  1195. }
  1196. return ESP_OK;
  1197. }
  1198. esp_err_t rmt_get_counter_clock(rmt_channel_t channel, uint32_t *clock_hz)
  1199. {
  1200. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1201. RMT_CHECK(clock_hz, "parameter clock_hz can't be null", ESP_ERR_INVALID_ARG);
  1202. RMT_ENTER_CRITICAL();
  1203. uint32_t rmt_source_clk_hz = 0;
  1204. #if SOC_RMT_SOURCE_CLK_INDEPENDENT
  1205. rmt_source_clk_hz = s_rmt_source_clock_hz[channel];
  1206. #else
  1207. rmt_source_clk_hz = s_rmt_source_clock_hz;
  1208. #endif
  1209. if (RMT_IS_RX_CHANNEL(channel)) {
  1210. *clock_hz = rmt_source_clk_hz / rmt_ll_rx_get_counter_clock_div(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  1211. } else {
  1212. *clock_hz = rmt_source_clk_hz / rmt_ll_tx_get_counter_clock_div(rmt_contex.hal.regs, channel);
  1213. }
  1214. RMT_EXIT_CRITICAL();
  1215. return ESP_OK;
  1216. }
  1217. #if SOC_RMT_SUPPORT_TX_GROUP
  1218. esp_err_t rmt_add_channel_to_group(rmt_channel_t channel)
  1219. {
  1220. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1221. RMT_ENTER_CRITICAL();
  1222. rmt_ll_tx_enable_sync(rmt_contex.hal.regs, true);
  1223. rmt_ll_tx_add_channel_to_group(rmt_contex.hal.regs, channel);
  1224. rmt_ll_tx_reset_counter_clock_div(rmt_contex.hal.regs, channel);
  1225. RMT_EXIT_CRITICAL();
  1226. return ESP_OK;
  1227. }
  1228. esp_err_t rmt_remove_channel_from_group(rmt_channel_t channel)
  1229. {
  1230. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1231. RMT_ENTER_CRITICAL();
  1232. if (rmt_ll_tx_remove_channel_from_group(rmt_contex.hal.regs, channel) == 0) {
  1233. rmt_ll_tx_enable_sync(rmt_contex.hal.regs, false);
  1234. }
  1235. RMT_EXIT_CRITICAL();
  1236. return ESP_OK;
  1237. }
  1238. esp_err_t rmt_memory_rw_rst(rmt_channel_t channel)
  1239. {
  1240. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1241. RMT_ENTER_CRITICAL();
  1242. if (RMT_IS_RX_CHANNEL(channel)) {
  1243. rmt_ll_rx_reset_pointer(rmt_contex.hal.regs, RMT_DECODE_RX_CHANNEL(channel));
  1244. } else {
  1245. rmt_ll_tx_reset_pointer(rmt_contex.hal.regs, channel);
  1246. }
  1247. RMT_EXIT_CRITICAL();
  1248. return ESP_OK;
  1249. }
  1250. #endif
  1251. #if SOC_RMT_SUPPORT_TX_LOOP_COUNT
  1252. esp_err_t rmt_set_tx_loop_count(rmt_channel_t channel, uint32_t count)
  1253. {
  1254. RMT_CHECK(RMT_IS_TX_CHANNEL(channel), RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  1255. RMT_ENTER_CRITICAL();
  1256. rmt_ll_tx_set_loop_count(rmt_contex.hal.regs, channel, count);
  1257. RMT_EXIT_CRITICAL();
  1258. return ESP_OK;
  1259. }
  1260. #endif