rmt.c 27 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717
  1. // Copyright 2015-2016 Espressif Systems (Shanghai) PTE LTD
  2. //
  3. // Licensed under the Apache License, Version 2.0 (the "License");
  4. // you may not use this file except in compliance with the License.
  5. // You may obtain a copy of the License at
  6. // http://www.apache.org/licenses/LICENSE-2.0
  7. //
  8. // Unless required by applicable law or agreed to in writing, software
  9. // distributed under the License is distributed on an "AS IS" BASIS,
  10. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  11. // See the License for the specific language governing permissions and
  12. // limitations under the License.
  13. #include <esp_types.h>
  14. #include <string.h>
  15. #include <stdlib.h>
  16. #include "freertos/FreeRTOS.h"
  17. #include "freertos/semphr.h"
  18. #include "freertos/xtensa_api.h"
  19. #include "freertos/ringbuf.h"
  20. #include "esp_intr.h"
  21. #include "esp_log.h"
  22. #include "esp_err.h"
  23. #include "soc/gpio_sig_map.h"
  24. #include "soc/rmt_struct.h"
  25. #include "driver/periph_ctrl.h"
  26. #include "driver/rmt.h"
  27. #define RMT_SOUCCE_CLK_APB (APB_CLK_FREQ) /*!< RMT source clock is APB_CLK */
  28. #define RMT_SOURCE_CLK_REF (1 * 1000000) /*!< not used yet */
  29. #define RMT_SOURCE_CLK(select) ((select == RMT_BASECLK_REF) ? (RMT_SOURCE_CLK_REF) : (RMT_SOUCCE_CLK_APB)) /*! RMT source clock frequency */
  30. #define RMT_CHANNEL_ERROR_STR "RMT CHANNEL ERR"
  31. #define RMT_ADDR_ERROR_STR "RMT ADDRESS ERR"
  32. #define RMT_MEM_CNT_ERROR_STR "RMT MEM BLOCK NUM ERR"
  33. #define RMT_CARRIER_ERROR_STR "RMT CARRIER LEVEL ERR"
  34. #define RMT_MEM_OWNER_ERROR_STR "RMT MEM OWNER_ERR"
  35. #define RMT_BASECLK_ERROR_STR "RMT BASECLK ERR"
  36. #define RMT_WR_MEM_OVF_ERROR_STR "RMT WR MEM OVERFLOW"
  37. #define RMT_GPIO_ERROR_STR "RMT GPIO ERROR"
  38. #define RMT_MODE_ERROR_STR "RMT MODE ERROR"
  39. #define RMT_CLK_DIV_ERROR_STR "RMT CLK DIV ERR"
  40. #define RMT_DRIVER_ERROR_STR "RMT DRIVER ERR"
  41. #define RMT_DRIVER_LENGTH_ERROR_STR "RMT PARAM LEN ERROR"
  42. static const char* RMT_TAG = "RMT";
  43. static bool s_rmt_driver_installed = false;
  44. #define RMT_CHECK(a, str, ret) if (!(a)) { \
  45. ESP_LOGE(RMT_TAG,"%s:%d (%s):%s", __FILE__, __LINE__, __FUNCTION__, str); \
  46. return (ret); \
  47. }
  48. static portMUX_TYPE rmt_spinlock = portMUX_INITIALIZER_UNLOCKED;
  49. typedef struct {
  50. int tx_offset;
  51. int tx_len_rem;
  52. int tx_sub_len;
  53. rmt_channel_t channel;
  54. rmt_item32_t* tx_data;
  55. xSemaphoreHandle tx_sem;
  56. RingbufHandle_t tx_buf;
  57. RingbufHandle_t rx_buf;
  58. } rmt_obj_t;
  59. rmt_obj_t* p_rmt_obj[RMT_CHANNEL_MAX] = {0};
  60. static void rmt_set_tx_wrap_en(rmt_channel_t channel, bool en)
  61. {
  62. portENTER_CRITICAL(&rmt_spinlock);
  63. RMT.apb_conf.mem_tx_wrap_en = en;
  64. portEXIT_CRITICAL(&rmt_spinlock);
  65. }
  66. static void rmt_set_data_mode(rmt_data_mode_t data_mode)
  67. {
  68. portENTER_CRITICAL(&rmt_spinlock);
  69. RMT.apb_conf.fifo_mask = data_mode;
  70. portEXIT_CRITICAL(&rmt_spinlock);
  71. }
  72. esp_err_t rmt_set_clk_div(rmt_channel_t channel, uint8_t div_cnt)
  73. {
  74. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  75. RMT.conf_ch[channel].conf0.div_cnt = div_cnt;
  76. return ESP_OK;
  77. }
  78. esp_err_t rmt_get_clk_div(rmt_channel_t channel, uint8_t* div_cnt)
  79. {
  80. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  81. RMT_CHECK(div_cnt != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  82. *div_cnt = RMT.conf_ch[channel].conf0.div_cnt;
  83. return ESP_OK;
  84. }
  85. esp_err_t rmt_set_rx_idle_thresh(rmt_channel_t channel, uint16_t thresh)
  86. {
  87. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  88. RMT.conf_ch[channel].conf0.idle_thres = thresh;
  89. return ESP_OK;
  90. }
  91. esp_err_t rmt_get_rx_idle_thresh(rmt_channel_t channel, uint16_t *thresh)
  92. {
  93. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  94. RMT_CHECK(thresh != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  95. *thresh = RMT.conf_ch[channel].conf0.idle_thres;
  96. return ESP_OK;
  97. }
  98. esp_err_t rmt_set_mem_block_num(rmt_channel_t channel, uint8_t rmt_mem_num)
  99. {
  100. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  101. RMT_CHECK(rmt_mem_num < 16, RMT_MEM_CNT_ERROR_STR, ESP_ERR_INVALID_ARG);
  102. RMT.conf_ch[channel].conf0.mem_size = rmt_mem_num;
  103. return ESP_OK;
  104. }
  105. esp_err_t rmt_get_mem_block_num(rmt_channel_t channel, uint8_t* rmt_mem_num)
  106. {
  107. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  108. RMT_CHECK(rmt_mem_num != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  109. *rmt_mem_num = RMT.conf_ch[channel].conf0.mem_size;
  110. return ESP_OK;
  111. }
  112. esp_err_t rmt_set_tx_carrier(rmt_channel_t channel, bool carrier_en, uint16_t high_level, uint16_t low_level,
  113. rmt_carrier_level_t carrier_level)
  114. {
  115. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  116. RMT_CHECK(carrier_level < RMT_CARRIER_LEVEL_MAX, RMT_CARRIER_ERROR_STR, ESP_ERR_INVALID_ARG);
  117. RMT.carrier_duty_ch[channel].high = high_level;
  118. RMT.carrier_duty_ch[channel].low = low_level;
  119. RMT.conf_ch[channel].conf0.carrier_out_lv = carrier_level;
  120. RMT.conf_ch[channel].conf0.carrier_en = carrier_en;
  121. return ESP_OK;
  122. }
  123. esp_err_t rmt_set_mem_pd(rmt_channel_t channel, bool pd_en)
  124. {
  125. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  126. RMT.conf_ch[channel].conf0.mem_pd = pd_en;
  127. return ESP_OK;
  128. }
  129. esp_err_t rmt_get_mem_pd(rmt_channel_t channel, bool* pd_en)
  130. {
  131. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  132. *pd_en = (bool) RMT.conf_ch[channel].conf0.mem_pd;
  133. return ESP_OK;
  134. }
  135. esp_err_t rmt_tx_start(rmt_channel_t channel, bool tx_idx_rst)
  136. {
  137. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  138. portENTER_CRITICAL(&rmt_spinlock);
  139. if(tx_idx_rst) {
  140. RMT.conf_ch[channel].conf1.mem_rd_rst = 1;
  141. }
  142. RMT.conf_ch[channel].conf1.mem_owner = RMT_MEM_OWNER_TX;
  143. RMT.conf_ch[channel].conf1.tx_start = 1;
  144. portEXIT_CRITICAL(&rmt_spinlock);
  145. return ESP_OK;
  146. }
  147. esp_err_t rmt_tx_stop(rmt_channel_t channel)
  148. {
  149. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  150. portENTER_CRITICAL(&rmt_spinlock);
  151. RMT.conf_ch[channel].conf1.tx_start = 0;
  152. portEXIT_CRITICAL(&rmt_spinlock);
  153. return ESP_OK;
  154. }
  155. esp_err_t rmt_rx_start(rmt_channel_t channel, bool rx_idx_rst)
  156. {
  157. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  158. portENTER_CRITICAL(&rmt_spinlock);
  159. if(rx_idx_rst) {
  160. RMT.conf_ch[channel].conf1.mem_wr_rst = 1;
  161. }
  162. RMT.conf_ch[channel].conf1.rx_en = 0;
  163. RMT.conf_ch[channel].conf1.mem_owner = RMT_MEM_OWNER_RX;
  164. RMT.conf_ch[channel].conf1.rx_en = 1;
  165. portEXIT_CRITICAL(&rmt_spinlock);
  166. return ESP_OK;
  167. }
  168. esp_err_t rmt_rx_stop(rmt_channel_t channel)
  169. {
  170. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  171. portENTER_CRITICAL(&rmt_spinlock);
  172. RMT.conf_ch[channel].conf1.rx_en = 0;
  173. portEXIT_CRITICAL(&rmt_spinlock);
  174. return ESP_OK;
  175. }
  176. esp_err_t rmt_memory_rw_rst(rmt_channel_t channel)
  177. {
  178. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  179. portENTER_CRITICAL(&rmt_spinlock);
  180. RMT.conf_ch[channel].conf1.mem_rd_rst = 1;
  181. RMT.conf_ch[channel].conf1.mem_wr_rst = 1;
  182. portEXIT_CRITICAL(&rmt_spinlock);
  183. return ESP_OK;
  184. }
  185. esp_err_t rmt_set_memory_owner(rmt_channel_t channel, rmt_mem_owner_t owner)
  186. {
  187. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  188. RMT_CHECK(owner < RMT_MEM_OWNER_MAX, RMT_MEM_OWNER_ERROR_STR, ESP_ERR_INVALID_ARG);
  189. portENTER_CRITICAL(&rmt_spinlock);
  190. RMT.conf_ch[channel].conf1.mem_owner = owner;
  191. portEXIT_CRITICAL(&rmt_spinlock);
  192. return ESP_OK;
  193. }
  194. esp_err_t rmt_get_memory_owner(rmt_channel_t channel, rmt_mem_owner_t* owner)
  195. {
  196. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  197. RMT_CHECK(owner != NULL, RMT_MEM_OWNER_ERROR_STR, ESP_ERR_INVALID_ARG);
  198. *owner = (rmt_mem_owner_t) RMT.conf_ch[channel].conf1.mem_owner;
  199. return ESP_OK;
  200. }
  201. esp_err_t rmt_set_tx_loop_mode(rmt_channel_t channel, bool loop_en)
  202. {
  203. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  204. portENTER_CRITICAL(&rmt_spinlock);
  205. RMT.conf_ch[channel].conf1.tx_conti_mode = loop_en;
  206. portEXIT_CRITICAL(&rmt_spinlock);
  207. return ESP_OK;
  208. }
  209. esp_err_t rmt_get_tx_loop_mode(rmt_channel_t channel, bool* loop_en)
  210. {
  211. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  212. *loop_en = (bool) RMT.conf_ch[channel].conf1.tx_conti_mode;
  213. return ESP_OK;
  214. }
  215. esp_err_t rmt_set_rx_filter(rmt_channel_t channel, bool rx_filter_en, uint8_t thresh)
  216. {
  217. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  218. portENTER_CRITICAL(&rmt_spinlock);
  219. RMT.conf_ch[channel].conf1.rx_filter_en = rx_filter_en;
  220. RMT.conf_ch[channel].conf1.rx_filter_thres = thresh;
  221. portEXIT_CRITICAL(&rmt_spinlock);
  222. return ESP_OK;
  223. }
  224. esp_err_t rmt_set_source_clk(rmt_channel_t channel, rmt_source_clk_t base_clk)
  225. {
  226. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  227. RMT_CHECK(base_clk < RMT_BASECLK_MAX, RMT_BASECLK_ERROR_STR, ESP_ERR_INVALID_ARG);
  228. portENTER_CRITICAL(&rmt_spinlock);
  229. RMT.conf_ch[channel].conf1.ref_always_on = base_clk;
  230. portEXIT_CRITICAL(&rmt_spinlock);
  231. return ESP_OK;
  232. }
  233. esp_err_t rmt_get_source_clk(rmt_channel_t channel, rmt_source_clk_t* src_clk)
  234. {
  235. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  236. *src_clk = (rmt_source_clk_t) (RMT.conf_ch[channel].conf1.ref_always_on);
  237. return ESP_OK;
  238. }
  239. esp_err_t rmt_set_idle_level(rmt_channel_t channel, bool idle_out_en, rmt_idle_level_t level)
  240. {
  241. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  242. RMT_CHECK(level < RMT_IDLE_LEVEL_MAX, "RMT IDLE LEVEL ERR", ESP_ERR_INVALID_ARG);
  243. portENTER_CRITICAL(&rmt_spinlock);
  244. RMT.conf_ch[channel].conf1.idle_out_en = idle_out_en;
  245. RMT.conf_ch[channel].conf1.idle_out_lv = level;
  246. portEXIT_CRITICAL(&rmt_spinlock);
  247. return ESP_OK;
  248. }
  249. esp_err_t rmt_get_status(rmt_channel_t channel, uint32_t* status)
  250. {
  251. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  252. *status = RMT.status_ch[channel];
  253. return ESP_OK;
  254. }
  255. rmt_data_mode_t rmt_get_data_mode()
  256. {
  257. return (rmt_data_mode_t) (RMT.apb_conf.fifo_mask);
  258. }
  259. void rmt_set_intr_enable_mask(uint32_t mask)
  260. {
  261. portENTER_CRITICAL(&rmt_spinlock);
  262. RMT.int_ena.val |= mask;
  263. portEXIT_CRITICAL(&rmt_spinlock);
  264. }
  265. void rmt_clr_intr_enable_mask(uint32_t mask)
  266. {
  267. portENTER_CRITICAL(&rmt_spinlock);
  268. RMT.int_ena.val &= (~mask);
  269. portEXIT_CRITICAL(&rmt_spinlock);
  270. }
  271. esp_err_t rmt_set_rx_intr_en(rmt_channel_t channel, bool en)
  272. {
  273. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  274. if(en) {
  275. rmt_set_intr_enable_mask(BIT(channel * 3 + 1));
  276. } else {
  277. rmt_clr_intr_enable_mask(BIT(channel * 3 + 1));
  278. }
  279. return ESP_OK;
  280. }
  281. esp_err_t rmt_set_err_intr_en(rmt_channel_t channel, bool en)
  282. {
  283. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  284. if(en) {
  285. rmt_set_intr_enable_mask(BIT(channel * 3 + 2));
  286. } else {
  287. rmt_clr_intr_enable_mask(BIT(channel * 3 + 2));
  288. }
  289. return ESP_OK;
  290. }
  291. esp_err_t rmt_set_tx_intr_en(rmt_channel_t channel, bool en)
  292. {
  293. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  294. if(en) {
  295. rmt_set_intr_enable_mask(BIT(channel * 3));
  296. } else {
  297. rmt_clr_intr_enable_mask(BIT(channel * 3));
  298. }
  299. return ESP_OK;
  300. }
  301. esp_err_t rmt_set_evt_intr_en(rmt_channel_t channel, bool en, uint16_t evt_thresh)
  302. {
  303. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  304. RMT_CHECK(evt_thresh < 256, "RMT EVT THRESH ERR", ESP_ERR_INVALID_ARG);
  305. if(en) {
  306. RMT.tx_lim_ch[channel].limit = evt_thresh;
  307. rmt_set_tx_wrap_en(channel, true);
  308. rmt_set_intr_enable_mask(BIT(channel + 24));
  309. } else {
  310. rmt_clr_intr_enable_mask(BIT(channel + 24));
  311. }
  312. return ESP_OK;
  313. }
  314. esp_err_t rmt_set_pin(rmt_channel_t channel, rmt_mode_t mode, gpio_num_t gpio_num)
  315. {
  316. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  317. RMT_CHECK(mode < RMT_MODE_MAX, RMT_MODE_ERROR_STR, ESP_ERR_INVALID_ARG);
  318. RMT_CHECK(((GPIO_IS_VALID_GPIO(gpio_num) && (mode == RMT_MODE_RX)) || (GPIO_IS_VALID_OUTPUT_GPIO(gpio_num) && (mode == RMT_MODE_TX))),
  319. RMT_GPIO_ERROR_STR, ESP_ERR_INVALID_ARG);
  320. PIN_FUNC_SELECT(GPIO_PIN_MUX_REG[gpio_num], 2);
  321. if(mode == RMT_MODE_TX) {
  322. gpio_set_direction(gpio_num, GPIO_MODE_OUTPUT);
  323. gpio_matrix_out(gpio_num, RMT_SIG_OUT0_IDX + channel, 0, 0);
  324. } else {
  325. gpio_set_direction(gpio_num, GPIO_MODE_INPUT);
  326. gpio_matrix_in(gpio_num, RMT_SIG_IN0_IDX + channel, 0);
  327. }
  328. return ESP_OK;
  329. }
  330. esp_err_t rmt_config(rmt_config_t* rmt_param)
  331. {
  332. uint8_t mode = rmt_param->rmt_mode;
  333. uint8_t channel = rmt_param->channel;
  334. uint8_t gpio_num = rmt_param->gpio_num;
  335. uint8_t mem_cnt = rmt_param->mem_block_num;
  336. int clk_div = rmt_param->clk_div;
  337. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  338. RMT_CHECK(GPIO_IS_VALID_GPIO(gpio_num), RMT_GPIO_ERROR_STR, ESP_ERR_INVALID_ARG);
  339. RMT_CHECK((mem_cnt + channel <= 8 && mem_cnt > 0), RMT_MEM_CNT_ERROR_STR, ESP_ERR_INVALID_ARG);
  340. RMT_CHECK((clk_div > 0), RMT_CLK_DIV_ERROR_STR, ESP_ERR_INVALID_ARG);
  341. periph_module_enable(PERIPH_RMT_MODULE);
  342. RMT.conf_ch[channel].conf0.div_cnt = clk_div;
  343. /*Visit data use memory not FIFO*/
  344. rmt_set_data_mode(RMT_DATA_MODE_MEM);
  345. /*Reset tx/rx memory index */
  346. portENTER_CRITICAL(&rmt_spinlock);
  347. RMT.conf_ch[channel].conf1.mem_rd_rst = 1;
  348. RMT.conf_ch[channel].conf1.mem_wr_rst = 1;
  349. portEXIT_CRITICAL(&rmt_spinlock);
  350. if(mode == RMT_MODE_TX) {
  351. uint32_t rmt_source_clk_hz = 0;
  352. uint32_t carrier_freq_hz = rmt_param->tx_config.carrier_freq_hz;
  353. uint16_t carrier_duty_percent = rmt_param->tx_config.carrier_duty_percent;
  354. uint8_t carrier_level = rmt_param->tx_config.carrier_level;
  355. uint8_t idle_level = rmt_param->tx_config.idle_level;
  356. portENTER_CRITICAL(&rmt_spinlock);
  357. RMT.conf_ch[channel].conf1.tx_conti_mode = rmt_param->tx_config.loop_en;
  358. /*Memory set block number*/
  359. RMT.conf_ch[channel].conf0.mem_size = mem_cnt;
  360. RMT.conf_ch[channel].conf1.mem_owner = RMT_MEM_OWNER_TX;
  361. /*We use APB clock in this version, which is 80Mhz, later we will release system reference clock*/
  362. RMT.conf_ch[channel].conf1.ref_always_on = RMT_BASECLK_APB;
  363. rmt_source_clk_hz = RMT_SOURCE_CLK(RMT_BASECLK_APB);
  364. /*Set idle level */
  365. RMT.conf_ch[channel].conf1.idle_out_en = rmt_param->tx_config.idle_output_en;
  366. RMT.conf_ch[channel].conf1.idle_out_lv = idle_level;
  367. portEXIT_CRITICAL(&rmt_spinlock);
  368. /*Set carrier*/
  369. uint32_t duty_div, duty_h, duty_l;
  370. duty_div = rmt_source_clk_hz / carrier_freq_hz;
  371. duty_h = duty_div * carrier_duty_percent / 100;
  372. duty_l = duty_div - duty_h;
  373. RMT.conf_ch[channel].conf0.carrier_out_lv = carrier_level;
  374. RMT.carrier_duty_ch[channel].high = duty_h;
  375. RMT.carrier_duty_ch[channel].low = duty_l;
  376. RMT.conf_ch[channel].conf0.carrier_en = rmt_param->tx_config.carrier_en;
  377. ESP_LOGD(RMT_TAG, "Rmt Tx Channel %u|Gpio %u|Sclk_Hz %u|Div %u|Carrier_Hz %u|Duty %u",
  378. channel, gpio_num, rmt_source_clk_hz, clk_div, carrier_freq_hz, carrier_duty_percent);
  379. }
  380. else if(RMT_MODE_RX == mode) {
  381. uint8_t filter_cnt = rmt_param->rx_config.filter_ticks_thresh;
  382. uint16_t threshold = rmt_param->rx_config.idle_threshold;
  383. portENTER_CRITICAL(&rmt_spinlock);
  384. /*clock init*/
  385. RMT.conf_ch[channel].conf1.ref_always_on = RMT_BASECLK_APB;
  386. uint32_t rmt_source_clk_hz = RMT_SOURCE_CLK(RMT_BASECLK_APB);
  387. /*memory set block number and owner*/
  388. RMT.conf_ch[channel].conf0.mem_size = mem_cnt;
  389. RMT.conf_ch[channel].conf1.mem_owner = RMT_MEM_OWNER_RX;
  390. /*Set idle threshold*/
  391. RMT.conf_ch[channel].conf0.idle_thres = threshold;
  392. /* Set RX filter */
  393. RMT.conf_ch[channel].conf1.rx_filter_thres = filter_cnt;
  394. RMT.conf_ch[channel].conf1.rx_filter_en = rmt_param->rx_config.filter_en;
  395. portEXIT_CRITICAL(&rmt_spinlock);
  396. ESP_LOGD(RMT_TAG, "Rmt Rx Channel %u|Gpio %u|Sclk_Hz %u|Div %u|Thresold %u|Filter %u",
  397. channel, gpio_num, rmt_source_clk_hz, clk_div, threshold, filter_cnt);
  398. }
  399. rmt_set_pin(channel, mode, gpio_num);
  400. return ESP_OK;
  401. }
  402. static void IRAM_ATTR rmt_fill_memory(rmt_channel_t channel, rmt_item32_t* item, uint16_t item_num, uint16_t mem_offset)
  403. {
  404. portENTER_CRITICAL(&rmt_spinlock);
  405. RMT.apb_conf.fifo_mask = RMT_DATA_MODE_MEM;
  406. portEXIT_CRITICAL(&rmt_spinlock);
  407. int i;
  408. for(i = 0; i < item_num; i++) {
  409. RMTMEM.chan[channel].data32[i + mem_offset].val = item[i].val;
  410. }
  411. }
  412. esp_err_t rmt_fill_tx_items(rmt_channel_t channel, rmt_item32_t* item, uint16_t item_num, uint16_t mem_offset)
  413. {
  414. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, (0));
  415. RMT_CHECK((item != NULL), RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  416. RMT_CHECK((item_num > 0), RMT_DRIVER_LENGTH_ERROR_STR, ESP_ERR_INVALID_ARG);
  417. /*Each block has 64 x 32 bits of data*/
  418. uint8_t mem_cnt = RMT.conf_ch[channel].conf0.mem_size;
  419. RMT_CHECK((mem_cnt * RMT_MEM_ITEM_NUM >= item_num), RMT_WR_MEM_OVF_ERROR_STR, ESP_ERR_INVALID_ARG);
  420. rmt_fill_memory(channel, item, item_num, mem_offset);
  421. return ESP_OK;
  422. }
  423. esp_err_t rmt_isr_register(uint8_t rmt_intr_num, void (*fn)(void*), void * arg)
  424. {
  425. RMT_CHECK((fn != NULL), RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  426. RMT_CHECK(s_rmt_driver_installed == false, "RMT DRIVER INSTALLED, CAN NOT REG ISR HANDLER", ESP_FAIL);
  427. portENTER_CRITICAL(&rmt_spinlock);
  428. ESP_INTR_DISABLE(rmt_intr_num);
  429. intr_matrix_set(xPortGetCoreID(), ETS_RMT_INTR_SOURCE, rmt_intr_num);
  430. xt_set_interrupt_handler(rmt_intr_num, fn, arg);
  431. ESP_INTR_ENABLE(rmt_intr_num);
  432. portEXIT_CRITICAL(&rmt_spinlock);
  433. return ESP_OK;
  434. }
  435. static int IRAM_ATTR rmt_get_mem_len(rmt_channel_t channel)
  436. {
  437. int block_num = RMT.conf_ch[channel].conf0.mem_size;
  438. int item_block_len = block_num * RMT_MEM_ITEM_NUM;
  439. volatile rmt_item32_t* data = RMTMEM.chan[channel].data32;
  440. int idx;
  441. for(idx = 0; idx < item_block_len; idx++) {
  442. if(data[idx].duration0 == 0) {
  443. return idx;
  444. } else if(data[idx].duration1 == 0) {
  445. return idx + 1;
  446. }
  447. }
  448. return idx;
  449. }
  450. static void IRAM_ATTR rmt_driver_isr_default(void* arg)
  451. {
  452. uint32_t intr_st = RMT.int_st.val;
  453. uint32_t i = 0;
  454. uint8_t channel;
  455. portBASE_TYPE HPTaskAwoken = 0;
  456. for(i = 0; i < 32; i++) {
  457. if(i < 24) {
  458. if(intr_st & (BIT(i))) {
  459. channel = i / 3;
  460. rmt_obj_t* p_rmt = p_rmt_obj[channel];
  461. switch(i % 3) {
  462. //TX END
  463. case 0:
  464. ESP_EARLY_LOGD(RMT_TAG, "RMT INTR : TX END\n");
  465. xSemaphoreGiveFromISR(p_rmt->tx_sem, &HPTaskAwoken);
  466. if(HPTaskAwoken == pdTRUE) {
  467. portYIELD_FROM_ISR();
  468. }
  469. p_rmt->tx_data = NULL;
  470. p_rmt->tx_len_rem = 0;
  471. p_rmt->tx_offset = 0;
  472. p_rmt->tx_sub_len = 0;
  473. break;
  474. //RX_END
  475. case 1:
  476. ESP_EARLY_LOGD(RMT_TAG, "RMT INTR : RX END");
  477. RMT.conf_ch[channel].conf1.rx_en = 0;
  478. int item_len = rmt_get_mem_len(channel);
  479. //change memory owner to protect data.
  480. RMT.conf_ch[channel].conf1.mem_owner = RMT_MEM_OWNER_TX;
  481. if(p_rmt->rx_buf) {
  482. BaseType_t res = xRingbufferSendFromISR(p_rmt->rx_buf, (void*) RMTMEM.chan[channel].data32, item_len * 4, &HPTaskAwoken);
  483. if(res == pdFALSE) {
  484. ESP_LOGE(RMT_TAG, "RMT RX BUFFER FULL");
  485. } else {
  486. }
  487. if(HPTaskAwoken == pdTRUE) {
  488. portYIELD_FROM_ISR();
  489. }
  490. } else {
  491. ESP_EARLY_LOGE(RMT_TAG, "RMT RX BUFFER ERROR\n");
  492. }
  493. RMT.conf_ch[channel].conf1.mem_wr_rst = 1;
  494. RMT.conf_ch[channel].conf1.mem_owner = RMT_MEM_OWNER_RX;
  495. RMT.conf_ch[channel].conf1.rx_en = 1;
  496. break;
  497. //ERR
  498. case 2:
  499. ESP_EARLY_LOGE(RMT_TAG, "RMT[%d] ERR", channel);
  500. ESP_EARLY_LOGE(RMT_TAG, "status: 0x%08x", RMT.status_ch[channel]);
  501. RMT.int_ena.val &= (~(BIT(i)));
  502. break;
  503. default:
  504. break;
  505. }
  506. RMT.int_clr.val = BIT(i);
  507. }
  508. } else {
  509. if(intr_st & (BIT(i))) {
  510. channel = i - 24;
  511. rmt_obj_t* p_rmt = p_rmt_obj[channel];
  512. RMT.int_clr.val = BIT(i);
  513. ESP_EARLY_LOGD(RMT_TAG, "RMT CH[%d]: EVT INTR", channel);
  514. if(p_rmt->tx_data == NULL) {
  515. //skip
  516. } else {
  517. rmt_item32_t* pdata = p_rmt->tx_data;
  518. int len_rem = p_rmt->tx_len_rem;
  519. if(len_rem >= p_rmt->tx_sub_len) {
  520. rmt_fill_memory(channel, pdata, p_rmt->tx_sub_len, p_rmt->tx_offset);
  521. p_rmt->tx_data += p_rmt->tx_sub_len;
  522. p_rmt->tx_len_rem -= p_rmt->tx_sub_len;
  523. } else if(len_rem == 0) {
  524. RMTMEM.chan[channel].data32[p_rmt->tx_offset].val = 0;
  525. } else {
  526. rmt_fill_memory(channel, pdata, len_rem, p_rmt->tx_offset);
  527. RMTMEM.chan[channel].data32[p_rmt->tx_offset + len_rem].val = 0;
  528. p_rmt->tx_data += len_rem;
  529. p_rmt->tx_len_rem -= len_rem;
  530. }
  531. if(p_rmt->tx_offset == 0) {
  532. p_rmt->tx_offset = p_rmt->tx_sub_len;
  533. } else {
  534. p_rmt->tx_offset = 0;
  535. }
  536. }
  537. }
  538. }
  539. }
  540. }
  541. esp_err_t rmt_driver_uninstall(rmt_channel_t channel)
  542. {
  543. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  544. if(p_rmt_obj[channel] == NULL) {
  545. return ESP_OK;
  546. }
  547. xSemaphoreTake(p_rmt_obj[channel]->tx_sem, portMAX_DELAY);
  548. rmt_set_rx_intr_en(channel, 0);
  549. rmt_set_err_intr_en(channel, 0);
  550. rmt_set_tx_intr_en(channel, 0);
  551. rmt_set_evt_intr_en(channel, 0, 0xffff);
  552. if(p_rmt_obj[channel]->tx_sem) {
  553. vSemaphoreDelete(p_rmt_obj[channel]->tx_sem);
  554. p_rmt_obj[channel]->tx_sem = NULL;
  555. }
  556. if(p_rmt_obj[channel]->rx_buf) {
  557. vRingbufferDelete(p_rmt_obj[channel]->rx_buf);
  558. p_rmt_obj[channel]->rx_buf = NULL;
  559. }
  560. free(p_rmt_obj[channel]);
  561. p_rmt_obj[channel] = NULL;
  562. s_rmt_driver_installed = false;
  563. return ESP_OK;
  564. }
  565. esp_err_t rmt_driver_install(rmt_channel_t channel, size_t rx_buf_size, int rmt_intr_num)
  566. {
  567. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  568. if(p_rmt_obj[channel] != NULL) {
  569. ESP_LOGD(RMT_TAG, "RMT DRIVER ALREADY INSTALLED");
  570. return ESP_FAIL;
  571. }
  572. ESP_INTR_DISABLE(rmt_intr_num);
  573. p_rmt_obj[channel] = (rmt_obj_t*) malloc(sizeof(rmt_obj_t));
  574. if(p_rmt_obj[channel] == NULL) {
  575. ESP_LOGE(RMT_TAG, "RMT driver malloc error");
  576. return ESP_FAIL;
  577. }
  578. memset(p_rmt_obj[channel], 0, sizeof(rmt_obj_t));
  579. p_rmt_obj[channel]->tx_len_rem = 0;
  580. p_rmt_obj[channel]->tx_data = NULL;
  581. p_rmt_obj[channel]->channel = channel;
  582. p_rmt_obj[channel]->tx_offset = 0;
  583. p_rmt_obj[channel]->tx_sub_len = 0;
  584. if(p_rmt_obj[channel]->tx_sem == NULL) {
  585. p_rmt_obj[channel]->tx_sem = xSemaphoreCreateBinary();
  586. xSemaphoreGive(p_rmt_obj[channel]->tx_sem);
  587. }
  588. if(p_rmt_obj[channel]->rx_buf == NULL && rx_buf_size > 0) {
  589. p_rmt_obj[channel]->rx_buf = xRingbufferCreate(rx_buf_size, RINGBUF_TYPE_NOSPLIT);
  590. rmt_set_rx_intr_en(channel, 1);
  591. rmt_set_err_intr_en(channel, 1);
  592. }
  593. if(s_rmt_driver_installed == false) {
  594. rmt_isr_register(rmt_intr_num, rmt_driver_isr_default, NULL);
  595. s_rmt_driver_installed = true;
  596. }
  597. rmt_set_tx_intr_en(channel, 1);
  598. ESP_INTR_ENABLE(rmt_intr_num);
  599. return ESP_OK;
  600. }
  601. esp_err_t rmt_write_items(rmt_channel_t channel, rmt_item32_t* rmt_item, int item_num, bool wait_tx_done)
  602. {
  603. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  604. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  605. RMT_CHECK(rmt_item != NULL, RMT_ADDR_ERROR_STR, ESP_FAIL);
  606. RMT_CHECK(item_num > 0, RMT_DRIVER_LENGTH_ERROR_STR, ESP_ERR_INVALID_ARG);
  607. rmt_obj_t* p_rmt = p_rmt_obj[channel];
  608. int block_num = RMT.conf_ch[channel].conf0.mem_size;
  609. int item_block_len = block_num * RMT_MEM_ITEM_NUM;
  610. int item_sub_len = block_num * RMT_MEM_ITEM_NUM / 2;
  611. int len_rem = item_num;
  612. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  613. // fill the memory block first
  614. if(item_num >= item_block_len) {
  615. rmt_fill_memory(channel, rmt_item, item_block_len, 0);
  616. RMT.tx_lim_ch[channel].limit = item_sub_len;
  617. RMT.apb_conf.mem_tx_wrap_en = 1;
  618. len_rem -= item_block_len;
  619. RMT.conf_ch[channel].conf1.tx_conti_mode = 0;
  620. rmt_set_evt_intr_en(channel, 1, item_sub_len);
  621. p_rmt->tx_data = rmt_item + item_block_len;
  622. p_rmt->tx_len_rem = len_rem;
  623. p_rmt->tx_offset = 0;
  624. p_rmt->tx_sub_len = item_sub_len;
  625. } else {
  626. rmt_fill_memory(channel, rmt_item, len_rem, 0);
  627. RMTMEM.chan[channel].data32[len_rem].val = 0;
  628. len_rem = 0;
  629. }
  630. rmt_tx_start(channel, true);
  631. if(wait_tx_done) {
  632. xSemaphoreTake(p_rmt->tx_sem, portMAX_DELAY);
  633. xSemaphoreGive(p_rmt->tx_sem);
  634. }
  635. return ESP_OK;
  636. }
  637. esp_err_t rmt_wait_tx_done(rmt_channel_t channel)
  638. {
  639. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  640. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  641. xSemaphoreTake(p_rmt_obj[channel]->tx_sem, portMAX_DELAY);
  642. xSemaphoreGive(p_rmt_obj[channel]->tx_sem);
  643. return ESP_OK;
  644. }
  645. esp_err_t rmt_get_ringbuf_handler(rmt_channel_t channel, RingbufHandle_t* buf_handler)
  646. {
  647. RMT_CHECK(channel < RMT_CHANNEL_MAX, RMT_CHANNEL_ERROR_STR, ESP_ERR_INVALID_ARG);
  648. RMT_CHECK(p_rmt_obj[channel] != NULL, RMT_DRIVER_ERROR_STR, ESP_FAIL);
  649. RMT_CHECK(buf_handler != NULL, RMT_ADDR_ERROR_STR, ESP_ERR_INVALID_ARG);
  650. *buf_handler = p_rmt_obj[channel]->rx_buf;
  651. return ESP_OK;
  652. }