start_gcc.S 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714
  1. /*
  2. * Copyright (c) 2006-2018, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2013-07-05 Bernard the first version
  9. * 2018-11-22 Jesven in the interrupt context, use rt_scheduler_do_irq_switch checks
  10. * and switches to a new thread
  11. * 2024-01-16 huanghe restructure this code section following the aarch64 architectural style
  12. */
  13. #include "rtconfig.h"
  14. #define ARM_CPU_STACK_SIZE_OFFSET 12
  15. #define ARM_CPU_STACK_SIZE (1<<ARM_CPU_STACK_SIZE_OFFSET)
  16. .equ Mode_USR, 0x10
  17. .equ Mode_FIQ, 0x11
  18. .equ Mode_IRQ, 0x12
  19. .equ Mode_SVC, 0x13
  20. .equ Mode_ABT, 0x17
  21. .equ Mode_UND, 0x1B
  22. .equ Mode_SYS, 0x1F
  23. .equ I_Bit, 0x80 /* when I bit is set, IRQ is disabled */
  24. .equ F_Bit, 0x40 /* when F bit is set, FIQ is disabled */
  25. /**
  26. * @brief Get the physical address of the symbol
  27. *
  28. * @param reg is the register to store the physical address
  29. * @param symbol is symbol name
  30. * @param _pvoff is the offset between the physical address and the virtual address
  31. */
  32. .macro get_phy, reg, symbol, _pvoff
  33. ldr \reg, =\symbol
  34. add \reg, \_pvoff
  35. .endm
  36. /**
  37. * @brief Calculate the offset between the physical address and the virtual address of the "_reset"
  38. *
  39. * @param tmp is the register which will be used to store the virtual address of the "_reset"
  40. * @param out is the register which will be used to store the pv_off (paddr - vaddr)
  41. */
  42. .macro get_pvoff, tmp, out
  43. ldr \tmp, =_reset
  44. adr \out, _reset
  45. sub \out, \out, \tmp
  46. .endm
  47. pv_off .req r11 /* Used to store the offset between physical address and the virtual address */
  48. cpu_id .req r10 /* Used to store the cpu id */
  49. /* reset entry */
  50. .globl _reset
  51. _reset:
  52. /* Calculate the offset between the physical address and the virtual address */
  53. get_pvoff r0, pv_off
  54. /* exit hyp mode */
  55. bl init_cpu_mode
  56. /* clear bss section */
  57. bl init_kernel_bss
  58. /* Initializes the assembly environment stack */
  59. bl init_cpu_stack_early
  60. /* init mmu */
  61. b init_mmu_page_table_early
  62. init_cpu_stack_early:
  63. cps #Mode_SVC
  64. get_phy r0, svc_stack_top, pv_off
  65. mov sp, r0
  66. #ifdef RT_USING_FPU
  67. mov r4, #0xfffffff
  68. mcr p15, 0, r4, c1, c0, 2 /* CPACR */
  69. #endif
  70. mov pc, lr
  71. init_kernel_bss:
  72. /* enable I cache + branch prediction */
  73. mrc p15, 0, r0, c1, c0, 0 /* SCTLR */
  74. orr r0, r0, #(1<<12) /* I=1 */
  75. orr r0, r0, #(1<<11) /* Z=1 */
  76. mcr p15, 0, r0, c1, c0, 0
  77. mov r0,#0
  78. get_phy r1, __bss_start, pv_off
  79. get_phy r2, __bss_end, pv_off
  80. bss_loop:
  81. cmp r1,r2 /* check if data to clear */
  82. strlo r0,[r1],#4 /* clear 4 bytes */
  83. blo bss_loop /* loop until done */
  84. mov pc, lr
  85. init_cpu_mode:
  86. #ifdef ARCH_ARMV8
  87. /* Check for HYP mode */
  88. mrs r0, cpsr_all
  89. and r0, r0, #0x1F
  90. mov r8, #0x1A
  91. cmp r0, r8
  92. beq overHyped
  93. b continue_exit
  94. overHyped: /* Get out of HYP mode */
  95. mov r9, lr
  96. /* HYP mode has a dedicated register, called ELR_hyp,
  97. to store the exception return address.
  98. The lr register needs to be temporarily saved,
  99. otherwise "mov pc lr" cannot be used after switching modes. */
  100. adr r1, continue_exit
  101. msr ELR_hyp, r1
  102. mrs r1, cpsr_all
  103. and r1, r1, #0xFFFFFFE0 /* CPSR_MODE_MASK */
  104. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  105. msr SPSR_hyp, r1
  106. eret
  107. continue_exit:
  108. mov lr ,r9
  109. #endif
  110. #ifdef SOC_BCM283x
  111. /* Suspend the other cpu cores */
  112. mrc p15, 0, r0, c0, c0, 5 /* MPIDR */
  113. ands r0, #3
  114. bne _halt
  115. /* Disable IRQ & FIQ */
  116. cpsid if
  117. /* Check for HYP mode */
  118. mrs r0, cpsr_all
  119. and r0, r0, #0x1F
  120. mov r8, #0x1A
  121. cmp r0, r8
  122. beq overHyped
  123. b continue_exit
  124. overHyped: /* Get out of HYP mode */
  125. mov r9, lr
  126. /* HYP mode has a dedicated register, called ELR_hyp,
  127. to store the exception return address.
  128. The lr register needs to be temporarily saved,
  129. otherwise "mov pc lr" cannot be used after switching modes. */
  130. adr r1, continue_exit
  131. msr ELR_hyp, r1
  132. mrs r1, cpsr_all
  133. and r1, r1, #0xFFFFFFE0 /* CPSR_MODE_MASK */
  134. orr r1, r1, #0x13 /* CPSR_MODE_SUPERVISOR */
  135. msr SPSR_hyp, r1
  136. eret
  137. continue_exit:
  138. mov lr ,r9
  139. /* set the cpu to SVC32 mode and disable interrupt */
  140. mrs r0, cpsr
  141. bic r0, r0, #0x1f
  142. orr r0, r0, #0x13
  143. msr cpsr_c, r0
  144. #endif
  145. /* disable MMU */
  146. mrc p15, 0, r0, c1, c0, 0 /* SCTLR */
  147. bic r0, #1 /* M=0 */
  148. mcr p15, 0, r0, c1, c0, 0
  149. dsb
  150. isb
  151. /* invalidate TLB, I-cache and branch predictor */
  152. mov r0, #0
  153. mcr p15, 0, r0, c8, c7, 0 /* ITLBIALL */
  154. mcr p15, 0, r0, c7, c5, 0 /* ICIALLU */
  155. mcr p15, 0, r0, c7, c5, 6 /* BPIALL */
  156. dsb
  157. isb
  158. mov pc, lr
  159. init_mmu_page_table_early:
  160. get_phy r0, init_mtbl, pv_off
  161. mov r1, pv_off
  162. bl rt_hw_mem_setup_early
  163. /* get cpu id */
  164. bl rt_hw_cpu_id_early
  165. mov cpu_id ,r0
  166. /* enable_mmu_page_table_early is changed to master_core_startup */
  167. ldr lr, =master_core_startup
  168. cmp cpu_id, #0
  169. beq enable_mmu_page_table_early
  170. #ifdef RT_USING_SMP
  171. #ifdef RT_SMP_AUTO_BOOT
  172. /* if cpu id > 0, stop or wait */
  173. ldr r0, =secondary_cpu_entry
  174. mov r1, #0
  175. str r1, [r0] /* clean secondary_cpu_entry */
  176. #endif
  177. #endif
  178. secondary_loop:
  179. @ cpu core 1 goes into sleep until core 0 wakeup it
  180. wfe
  181. #ifdef RT_SMP_AUTO_BOOT
  182. ldr r1, =secondary_cpu_entry
  183. ldr r0, [r1]
  184. cmp r0, #0
  185. blxne r0 /* if(secondary_cpu_entry) secondary_cpu_entry(); */
  186. #endif /* RT_SMP_AUTO_BOOT */
  187. b secondary_loop
  188. enable_mmu_page_table_early:
  189. /* init TTBR0 */
  190. get_phy r0, init_mtbl, pv_off
  191. mcr p15, #0, r0, c2, c0, #0
  192. dmb
  193. /* set all domains with client mode */
  194. ldr r0,=#0x55555555 /* client */
  195. mcr p15, #0, r0, c3, c0, #0 /* DACR */
  196. /* disable ttbr1 */
  197. mov r0, #(1 << 5) /* PD1=1 */
  198. mcr p15, 0, r0, c2, c0, 2 /* TTBCR */
  199. /* init stack for cpu mod */
  200. cps #Mode_UND
  201. ldr r1,=und_stack_top
  202. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  203. cps #Mode_IRQ
  204. ldr r1, =irq_stack_top
  205. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  206. cps #Mode_FIQ
  207. ldr r1, =irq_stack_top
  208. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  209. cps #Mode_ABT
  210. ldr r1, =abt_stack_top
  211. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  212. cps #Mode_SVC
  213. ldr r1, =svc_stack_top
  214. sub sp, r1, cpu_id, asl #ARM_CPU_STACK_SIZE_OFFSET
  215. /* invalidate TLB, I-cache and branch predictor */
  216. mov r0, #0
  217. mcr p15, 0, r0, c8, c7, 0 /* ITLBIALL */
  218. mcr p15, 0, r0, c7, c5, 0 /* ICIALLU */
  219. mcr p15, 0, r0, c7, c5, 6 /* BPIALL */
  220. /* enable I cache + branch prediction, enable MMU */
  221. mrc p15, 0, r0, c1, c0, 0 /* SCTLR */
  222. bic r0, r0, #0x7 /* C=0, A=0, M=0 */
  223. orr r0, #((1 << 12) | (1 << 11)) /* I=1, Z=1 */
  224. orr r0, #((1 << 2) | (1 << 0)) /* C=1, M=1 */
  225. mcr p15, 0, r0, c1, c0, 0
  226. dsb
  227. isb
  228. mov pc, lr
  229. master_core_startup :
  230. mov r0 ,pv_off
  231. bl rt_kmem_pvoff_set
  232. ldr lr, =rtthread_startup
  233. mov pc, lr
  234. .global rt_hw_mmu_tbl_get
  235. rt_hw_mmu_tbl_get:
  236. mrc p15, 0, r0, c2, c0, 0 /* TTBR0 */
  237. bic r0, #0x18 /* RGN=0 */
  238. mov pc, lr
  239. .weak rt_hw_cpu_id_early
  240. rt_hw_cpu_id_early:
  241. mrc p15, 0, r0, c0, c0, 5 /* MPIDR */
  242. and r0, r0, #0xf
  243. mov pc, lr
  244. #ifdef RT_USING_SMP
  245. .global rt_secondary_cpu_entry
  246. rt_secondary_cpu_entry:
  247. ldr r0, =_reset
  248. adr pv_off, _reset
  249. sub pv_off, pv_off, r0
  250. bl init_cpu_stack_early
  251. /* init mmu */
  252. bl rt_hw_cpu_id_early
  253. mov cpu_id ,r0
  254. ldr lr ,= rt_hw_secondary_cpu_bsp_start
  255. b enable_mmu_page_table_early
  256. #endif
  257. /* exception handlers: undef, swi, padt, dabt, resv, irq, fiq */
  258. .section .text.isr, "ax"
  259. .align 5
  260. .globl vector_fiq
  261. vector_fiq:
  262. stmfd sp!,{r0-r7,lr}
  263. bl rt_hw_trap_fiq
  264. ldmfd sp!,{r0-r7,lr}
  265. subs pc, lr, #4
  266. .globl rt_interrupt_enter
  267. .globl rt_interrupt_leave
  268. .globl rt_thread_switch_interrupt_flag
  269. .globl rt_interrupt_from_thread
  270. .globl rt_interrupt_to_thread
  271. .globl rt_current_thread
  272. .globl vmm_thread
  273. .globl vmm_virq_check
  274. .align 5
  275. .globl vector_irq
  276. vector_irq:
  277. #ifdef RT_USING_SMP
  278. stmfd sp!, {r0, r1}
  279. cps #Mode_SVC
  280. mov r0, sp /* svc_sp */
  281. mov r1, lr /* svc_lr */
  282. cps #Mode_IRQ
  283. sub lr, #4
  284. stmfd r0!, {r1, lr} /* svc_lr, svc_pc */
  285. stmfd r0!, {r2 - r12}
  286. ldmfd sp!, {r1, r2} /* original r0, r1 */
  287. stmfd r0!, {r1 - r2}
  288. mrs r1, spsr /* original mode */
  289. stmfd r0!, {r1}
  290. #ifdef RT_USING_SMART
  291. stmfd r0, {r13, r14}^ /* usr_sp, usr_lr */
  292. sub r0, #8
  293. #endif
  294. #ifdef RT_USING_FPU
  295. /* fpu context */
  296. vmrs r6, fpexc
  297. tst r6, #(1<<30)
  298. beq 1f
  299. vstmdb r0!, {d0-d15}
  300. vstmdb r0!, {d16-d31}
  301. vmrs r5, fpscr
  302. stmfd r0!, {r5}
  303. 1:
  304. stmfd r0!, {r6}
  305. #endif
  306. /* now irq stack is clean */
  307. /* r0 is task svc_sp */
  308. /* backup r0 -> r8 */
  309. mov r8, r0
  310. cps #Mode_SVC
  311. mov sp, r8
  312. bl rt_interrupt_enter
  313. bl rt_hw_trap_irq
  314. bl rt_interrupt_leave
  315. mov r0, r8
  316. bl rt_scheduler_do_irq_switch
  317. b rt_hw_context_switch_exit
  318. #else
  319. stmfd sp!, {r0-r12,lr}
  320. bl rt_interrupt_enter
  321. bl rt_hw_trap_irq
  322. bl rt_interrupt_leave
  323. /* if rt_thread_switch_interrupt_flag set, jump to
  324. * rt_hw_context_switch_interrupt_do and don't return */
  325. ldr r0, =rt_thread_switch_interrupt_flag
  326. ldr r1, [r0]
  327. cmp r1, #1
  328. beq rt_hw_context_switch_interrupt_do
  329. #ifdef RT_USING_SMART
  330. ldmfd sp!, {r0-r12,lr}
  331. cps #Mode_SVC
  332. push {r0-r12}
  333. mov r7, lr
  334. cps #Mode_IRQ
  335. mrs r4, spsr
  336. sub r5, lr, #4
  337. cps #Mode_SVC
  338. and r6, r4, #0x1f
  339. cmp r6, #0x10
  340. bne 1f
  341. msr spsr_csxf, r4
  342. mov lr, r5
  343. pop {r0-r12}
  344. b arch_ret_to_user
  345. 1:
  346. mov lr, r7
  347. cps #Mode_IRQ
  348. msr spsr_csxf, r4
  349. mov lr, r5
  350. cps #Mode_SVC
  351. pop {r0-r12}
  352. cps #Mode_IRQ
  353. movs pc, lr
  354. #else
  355. ldmfd sp!, {r0-r12,lr}
  356. subs pc, lr, #4
  357. #endif
  358. rt_hw_context_switch_interrupt_do:
  359. mov r1, #0 /* clear flag */
  360. str r1, [r0]
  361. mov r1, sp /* r1 point to {r0-r3} in stack */
  362. add sp, sp, #4*4
  363. ldmfd sp!, {r4-r12,lr} /* reload saved registers */
  364. mrs r0, spsr /* get cpsr of interrupt thread */
  365. sub r2, lr, #4 /* save old task's pc to r2 */
  366. /* Switch to SVC mode with no interrupt. If the usr mode guest is
  367. * interrupted, this will just switch to the stack of kernel space.
  368. * save the registers in kernel space won't trigger data abort. */
  369. msr cpsr_c, #I_Bit|F_Bit|Mode_SVC
  370. stmfd sp!, {r2} /* push old task's pc */
  371. stmfd sp!, {r4-r12,lr} /* push old task's lr,r12-r4 */
  372. ldmfd r1, {r1-r4} /* restore r0-r3 of the interrupt thread */
  373. stmfd sp!, {r1-r4} /* push old task's r0-r3 */
  374. stmfd sp!, {r0} /* push old task's cpsr */
  375. #ifdef RT_USING_SMART
  376. stmfd sp, {r13, r14}^ /*push usr_sp, usr_lr */
  377. sub sp, #8
  378. #endif
  379. #ifdef RT_USING_FPU
  380. /* fpu context */
  381. vmrs r6, fpexc
  382. tst r6, #(1<<30)
  383. beq 1f
  384. vstmdb sp!, {d0-d15}
  385. vstmdb sp!, {d16-d31}
  386. vmrs r5, fpscr
  387. stmfd sp!, {r5}
  388. 1:
  389. stmfd sp!, {r6}
  390. #endif
  391. ldr r4, =rt_interrupt_from_thread
  392. ldr r5, [r4]
  393. str sp, [r5] /* store sp in preempted tasks's TCB */
  394. ldr r6, =rt_interrupt_to_thread
  395. ldr r6, [r6]
  396. ldr sp, [r6] /* get new task's stack pointer */
  397. #ifdef RT_USING_SMART
  398. bl rt_thread_self
  399. mov r4, r0
  400. bl lwp_aspace_switch
  401. mov r0, r4
  402. bl lwp_user_setting_restore
  403. #endif
  404. #ifdef RT_USING_FPU
  405. /* fpu context */
  406. ldmfd sp!, {r6}
  407. vmsr fpexc, r6
  408. tst r6, #(1<<30)
  409. beq 1f
  410. ldmfd sp!, {r5}
  411. vmsr fpscr, r5
  412. vldmia sp!, {d16-d31}
  413. vldmia sp!, {d0-d15}
  414. 1:
  415. #endif
  416. #ifdef RT_USING_SMART
  417. ldmfd sp, {r13, r14}^ /*pop usr_sp, usr_lr */
  418. add sp, #8
  419. #endif
  420. ldmfd sp!, {r4} /* pop new task's cpsr to spsr */
  421. msr spsr_cxsf, r4
  422. #ifdef RT_USING_SMART
  423. and r4, #0x1f
  424. cmp r4, #0x10
  425. bne 1f
  426. ldmfd sp!, {r0-r12,lr}
  427. ldmfd sp!, {lr}
  428. b arch_ret_to_user
  429. 1:
  430. #endif
  431. /* pop new task's r0-r12,lr & pc, copy spsr to cpsr */
  432. ldmfd sp!, {r0-r12,lr,pc}^
  433. #endif
  434. .macro push_svc_reg
  435. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  436. stmia sp, {r0 - r12} /* Calling r0-r12 */
  437. mov r0, sp
  438. add sp, sp, #17 * 4
  439. mrs r6, spsr /* Save CPSR */
  440. str lr, [r0, #15*4] /* Push PC */
  441. str r6, [r0, #16*4] /* Push CPSR */
  442. and r1, r6, #0x1f
  443. cmp r1, #0x10
  444. cps #Mode_SYS
  445. streq sp, [r0, #13*4] /* Save calling SP */
  446. streq lr, [r0, #14*4] /* Save calling PC */
  447. cps #Mode_SVC
  448. strne sp, [r0, #13*4] /* Save calling SP */
  449. strne lr, [r0, #14*4] /* Save calling PC */
  450. .endm
  451. .align 5
  452. .weak vector_swi
  453. vector_swi:
  454. push_svc_reg
  455. bl rt_hw_trap_swi
  456. b .
  457. .align 5
  458. .globl vector_undef
  459. vector_undef:
  460. push_svc_reg
  461. bl rt_hw_trap_undef
  462. #ifdef RT_USING_FPU
  463. cps #Mode_UND
  464. sub sp, sp, #17 * 4
  465. ldr lr, [sp, #15*4]
  466. ldmia sp, {r0 - r12}
  467. add sp, sp, #17 * 4
  468. movs pc, lr
  469. #endif
  470. b .
  471. .align 5
  472. .globl vector_pabt
  473. vector_pabt:
  474. push_svc_reg
  475. #ifdef RT_USING_SMART
  476. /* cp Mode_ABT stack to SVC */
  477. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  478. mov lr, r0
  479. ldmia lr, {r0 - r12}
  480. stmia sp, {r0 - r12}
  481. add r1, lr, #13 * 4
  482. add r2, sp, #13 * 4
  483. ldmia r1, {r4 - r7}
  484. stmia r2, {r4 - r7}
  485. mov r0, sp
  486. bl rt_hw_trap_pabt
  487. /* return to user */
  488. ldr lr, [sp, #16*4] /* orign spsr */
  489. msr spsr_cxsf, lr
  490. ldr lr, [sp, #15*4] /* orign pc */
  491. ldmia sp, {r0 - r12}
  492. add sp, #17 * 4
  493. b arch_ret_to_user
  494. #else
  495. bl rt_hw_trap_pabt
  496. b .
  497. #endif
  498. .align 5
  499. .globl vector_dabt
  500. vector_dabt:
  501. push_svc_reg
  502. #ifdef RT_USING_SMART
  503. /* cp Mode_ABT stack to SVC */
  504. sub sp, sp, #17 * 4 /* Sizeof(struct rt_hw_exp_stack) */
  505. mov lr, r0
  506. ldmia lr, {r0 - r12}
  507. stmia sp, {r0 - r12}
  508. add r1, lr, #13 * 4
  509. add r2, sp, #13 * 4
  510. ldmia r1, {r4 - r7}
  511. stmia r2, {r4 - r7}
  512. mov r0, sp
  513. bl rt_hw_trap_dabt
  514. /* return to user */
  515. ldr lr, [sp, #16*4] /* orign spsr */
  516. msr spsr_cxsf, lr
  517. ldr lr, [sp, #15*4] /* orign pc */
  518. ldmia sp, {r0 - r12}
  519. add sp, #17 * 4
  520. b arch_ret_to_user
  521. #else
  522. bl rt_hw_trap_dabt
  523. b .
  524. #endif
  525. .align 5
  526. .globl vector_resv
  527. vector_resv:
  528. push_svc_reg
  529. bl rt_hw_trap_resv
  530. b .
  531. .global rt_hw_clz
  532. rt_hw_clz:
  533. clz r0, r0
  534. bx lr
  535. #include "asm-generic.h"
  536. START_POINT(_thread_start)
  537. mov r10, lr
  538. blx r1
  539. blx r10
  540. b . /* never here */
  541. START_POINT_END(_thread_start)
  542. .data
  543. .align 14
  544. init_mtbl:
  545. .space (4*4096) /* The L1 translation table therefore contains 4096 32-bit (word-sized) entries. */
  546. /*
  547. * void rt_hw_mmu_switch(rt_uint32_t* mmutable_p);
  548. * r0 --> mmutable_p (mmu table address)
  549. */
  550. .global rt_hw_mmu_switch
  551. rt_hw_mmu_switch:
  552. orr r0, #0x18 /* RGN=0b11 (Outer WB-WA) */
  553. mcr p15, 0, r0, c2, c0, 0 /* TTBR0 */
  554. /* invalidate TLB, I-cache and branch predictor */
  555. mov r0, #0
  556. mcr p15, 0, r0, c8, c7, 0 /* ITLBIALL */
  557. mcr p15, 0, r0, c7, c5, 0 /* ICIALLU */
  558. mcr p15, 0, r0, c7, c5, 6 /* BPIALL */
  559. dsb
  560. isb
  561. mov pc, lr
  562. .global rt_hw_set_process_id
  563. rt_hw_set_process_id:
  564. LSL r0, r0, #8
  565. MCR p15, 0, r0, c13, c0, 1 /* CONTEXTIDR */
  566. mov pc, lr
  567. .bss
  568. .align 3 /* align to 2~3=8 */
  569. .cpus_stack:
  570. svc_stack_n:
  571. #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
  572. .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
  573. #endif
  574. .space (ARM_CPU_STACK_SIZE)
  575. svc_stack_top:
  576. irq_stack_n:
  577. #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
  578. .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
  579. #endif
  580. .space (ARM_CPU_STACK_SIZE)
  581. irq_stack_top:
  582. und_stack_n:
  583. #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
  584. .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
  585. #endif
  586. .space (ARM_CPU_STACK_SIZE)
  587. und_stack_top:
  588. abt_stack_n:
  589. #if defined(RT_USING_SMP) && (RT_CPUS_NR > 1)
  590. .space ((RT_CPUS_NR - 1) * ARM_CPU_STACK_SIZE)
  591. #endif
  592. .space (ARM_CPU_STACK_SIZE)
  593. abt_stack_top: