/* * Copyright (c) 2006-2022, RT-Thread Development Team * * SPDX-License-Identifier: Apache-2.0 * * Change Logs: * Date Author Notes * 2018-10-06 ZhaoXiaowei the first version * 2021-11-04 GuEe-GUI set sp with SP_ELx * 2021-12-28 GuEe-GUI add fpu and smp support */ #include "rtconfig.h" #include "asm_fpu.h" #ifdef RT_USING_SMP #define rt_hw_interrupt_disable rt_hw_local_irq_disable #define rt_hw_interrupt_enable rt_hw_local_irq_enable #endif /* *enable gtimer */ .globl rt_hw_gtimer_enable rt_hw_gtimer_enable: MOV X0,#1 MSR CNTP_CTL_EL0,X0 RET /* *disable gtimer */ .globl rt_hw_gtimer_disable rt_hw_gtimer_disable: MSR CNTP_CTL_EL0,XZR RET /* *set gtimer CNTP_TVAL_EL0 value */ .globl rt_hw_set_gtimer_val rt_hw_set_gtimer_val: MSR CNTP_TVAL_EL0,X0 RET /* *get gtimer CNTP_TVAL_EL0 value */ .globl rt_hw_get_gtimer_val rt_hw_get_gtimer_val: MRS X0,CNTP_TVAL_EL0 RET .globl rt_hw_get_cntpct_val rt_hw_get_cntpct_val: MRS X0, CNTPCT_EL0 RET /* *get gtimer frq value */ .globl rt_hw_get_gtimer_frq rt_hw_get_gtimer_frq: MRS X0,CNTFRQ_EL0 RET /* *set gtimer frq value (only in EL3) */ .globl rt_hw_set_gtimer_frq rt_hw_set_gtimer_frq: MRS X1, CurrentEL CMP X1, 0xc BNE rt_hw_set_gtimer_frq_exit MSR CNTFRQ_EL0, X0 MOV X0, XZR rt_hw_set_gtimer_frq_exit: RET .macro SAVE_CONTEXT /* Save the entire context. */ SAVE_FPU SP STP X0, X1, [SP, #-0x10]! STP X2, X3, [SP, #-0x10]! STP X4, X5, [SP, #-0x10]! STP X6, X7, [SP, #-0x10]! STP X8, X9, [SP, #-0x10]! STP X10, X11, [SP, #-0x10]! STP X12, X13, [SP, #-0x10]! STP X14, X15, [SP, #-0x10]! STP X16, X17, [SP, #-0x10]! STP X18, X19, [SP, #-0x10]! STP X20, X21, [SP, #-0x10]! STP X22, X23, [SP, #-0x10]! STP X24, X25, [SP, #-0x10]! STP X26, X27, [SP, #-0x10]! STP X28, X29, [SP, #-0x10]! MRS X28, FPCR MRS X29, FPSR STP X28, X29, [SP, #-0x10]! STP X30, XZR, [SP, #-0x10]! MRS X0, CurrentEL CMP X0, 0xc B.EQ 3f CMP X0, 0x8 B.EQ 2f CMP X0, 0x4 B.EQ 1f B . 3: MRS X3, SPSR_EL3 /* Save the ELR. */ MRS X2, ELR_EL3 B 0f 2: MRS X3, SPSR_EL2 /* Save the ELR. */ MRS X2, ELR_EL2 B 0f 1: MRS X3, SPSR_EL1 MRS X2, ELR_EL1 B 0f 0: STP X2, X3, [SP, #-0x10]! MOV X0, SP /* Move SP into X0 for saving. */ .endm .macro SAVE_CONTEXT_T /* Save the entire context. */ SAVE_FPU SP STP X0, X1, [SP, #-0x10]! STP X2, X3, [SP, #-0x10]! STP X4, X5, [SP, #-0x10]! STP X6, X7, [SP, #-0x10]! STP X8, X9, [SP, #-0x10]! STP X10, X11, [SP, #-0x10]! STP X12, X13, [SP, #-0x10]! STP X14, X15, [SP, #-0x10]! STP X16, X17, [SP, #-0x10]! STP X18, X19, [SP, #-0x10]! STP X20, X21, [SP, #-0x10]! STP X22, X23, [SP, #-0x10]! STP X24, X25, [SP, #-0x10]! STP X26, X27, [SP, #-0x10]! STP X28, X29, [SP, #-0x10]! MRS X28, FPCR MRS X29, FPSR STP X28, X29, [SP, #-0x10]! STP X30, XZR, [SP, #-0x10]! MRS X0, CurrentEL CMP X0, 0xc B.EQ 3f CMP X0, 0x8 B.EQ 2f CMP X0, 0x4 B.EQ 1f B . 3: MOV X3, #((3 << 6) | 0x0d) /* EL3h */ MOV X2, X30 B 0f 2: MOV X3, #((3 << 6) | 0x09) /* EL2h */ MOV X2, X30 B 0f 1: MOV X3, #((3 << 6) | 0x05) /* EL1h */ MOV X2, X30 B 0f 0: STP X2, X3, [SP, #-0x10]! MOV X0, SP /* Move SP into X0 for saving. */ .endm .macro RESTORE_CONTEXT /* Set the SP to point to the stack of the task being restored. */ MOV SP, X0 LDP X2, X3, [SP], #0x10 /* SPSR and ELR. */ MRS X0, CurrentEL CMP X0, 0xc B.EQ 3f CMP X0, 0x8 B.EQ 2f CMP X0, 0x4 B.EQ 1f B . 3: MSR SPSR_EL3, X3 MSR ELR_EL3, X2 B 0f 2: MSR SPSR_EL2, X3 MSR ELR_EL2, X2 B 0f 1: MSR SPSR_EL1, X3 MSR ELR_EL1, X2 B 0f 0: LDP X30, XZR, [SP], #0x10 LDP X28, X29, [SP], #0x10 MSR FPCR, X28 MSR FPSR, X29 LDP X28, X29, [SP], #0x10 LDP X26, X27, [SP], #0x10 LDP X24, X25, [SP], #0x10 LDP X22, X23, [SP], #0x10 LDP X20, X21, [SP], #0x10 LDP X18, X19, [SP], #0x10 LDP X16, X17, [SP], #0x10 LDP X14, X15, [SP], #0x10 LDP X12, X13, [SP], #0x10 LDP X10, X11, [SP], #0x10 LDP X8, X9, [SP], #0x10 LDP X6, X7, [SP], #0x10 LDP X4, X5, [SP], #0x10 LDP X2, X3, [SP], #0x10 LDP X0, X1, [SP], #0x10 RESTORE_FPU SP ERET .endm .text /* * rt_base_t rt_hw_interrupt_disable(); */ .globl rt_hw_interrupt_disable rt_hw_interrupt_disable: MRS X0, DAIF MSR DAIFSet, #3 DSB SY RET /* * void rt_hw_interrupt_enable(rt_base_t level); */ .globl rt_hw_interrupt_enable rt_hw_interrupt_enable: DSB SY MOV X1, #0xC0 ANDS X0, X0, X1 B.NE rt_hw_interrupt_enable_exit MSR DAIFClr, #3 rt_hw_interrupt_enable_exit: RET /* * #ifdef RT_USING_SMP * void rt_hw_context_switch_to(rt_ubase_t to, stuct rt_thread *to_thread); * #else * void rt_hw_context_switch_to(rt_ubase_t to); * #endif * X0 --> to * X1 --> to_thread */ .globl rt_hw_context_switch_to rt_hw_context_switch_to: #ifdef RT_USING_SMP STR X0, [SP, #-0x8]! MOV X0, X1 BL rt_cpus_lock_status_restore LDR X0, [SP], #0x8 #endif /*RT_USING_SMP*/ LDR X0, [X0] RESTORE_CONTEXT .text /* * #ifdef RT_USING_SMP * void rt_hw_context_switch(rt_ubase_t from, rt_ubase_t to, struct rt_thread *to_thread); * #else * void rt_hw_context_switch(rt_ubase_t from, rt_ubase_t to); * #endif * X0 --> from * X1 --> to * X2 --> to_thread */ .globl rt_hw_context_switch rt_hw_context_switch: #ifdef RT_USING_SMP STP X0, X1, [SP, #-0x10]! STR X30, [SP, #-0x8]! MOV X0, X2 BL rt_cpus_lock_status_restore LDR X30, [SP], #0x8 LDP X0, X1, [SP], #0x10 #endif /*RT_USING_SMP*/ MOV X8,X0 MOV X9,X1 SAVE_CONTEXT_T STR X0, [X8] // store sp in preempted tasks TCB LDR X0, [X9] // get new task stack pointer RESTORE_CONTEXT /* * void rt_hw_context_switch_interrupt(rt_ubase_t from, rt_ubase_t to); */ .globl rt_thread_switch_interrupt_flag .globl rt_interrupt_from_thread .globl rt_interrupt_to_thread .globl rt_hw_context_switch_interrupt rt_hw_context_switch_interrupt: #ifdef RT_USING_SMP /* x0 = context */ /* x1 = ¤t_thread->sp */ /* x2 = &to_thread->sp, */ /* x3 = to_thread TCB */ STR X0, [X1] LDR X0, [x2] MOV SP, X0 MOV X0, X3 BL rt_cpus_lock_status_restore MOV X0, SP RESTORE_CONTEXT #else LDR X2, =rt_thread_switch_interrupt_flag LDR X3, [X2] CMP X3, #1 B.EQ _reswitch LDR X4, =rt_interrupt_from_thread // set rt_interrupt_from_thread MOV X3, #1 // set rt_thread_switch_interrupt_flag to 1 STR X0, [X4] STR X3, [X2] _reswitch: LDR X2, =rt_interrupt_to_thread // set rt_interrupt_to_thread STR X1, [X2] RET #endif .text // -- Exception handlers ---------------------------------- .align 8 .globl vector_fiq vector_fiq: SAVE_CONTEXT STP X0, X1, [SP, #-0x10]! BL rt_hw_trap_fiq LDP X0, X1, [SP], #0x10 RESTORE_CONTEXT .globl rt_interrupt_enter .globl rt_interrupt_leave .globl rt_thread_switch_interrupt_flag .globl rt_interrupt_from_thread .globl rt_interrupt_to_thread // ------------------------------------------------------------------- .align 8 .globl vector_irq vector_irq: SAVE_CONTEXT STP X0, X1, [SP, #-0x10]! BL rt_interrupt_enter BL rt_hw_trap_irq BL rt_interrupt_leave LDP X0, X1, [SP], #0x10 #ifdef RT_USING_SMP /* Never reture If can switch */ BL rt_scheduler_do_irq_switch MOV X0, SP #endif // if rt_thread_switch_interrupt_flag set, jump to // rt_hw_context_switch_interrupt_do and don't return LDR X1, =rt_thread_switch_interrupt_flag LDR X2, [X1] CMP X2, #1 B.NE vector_irq_exit MOV X2, #0 // clear flag STR X2, [X1] LDR X3, =rt_interrupt_from_thread LDR X4, [X3] STR x0, [X4] // store sp in preempted tasks's TCB LDR x3, =rt_interrupt_to_thread LDR X4, [X3] LDR x0, [X4] // get new task's stack pointer vector_irq_exit: RESTORE_CONTEXT // ------------------------------------------------- .align 8 .globl vector_error vector_error: SAVE_CONTEXT BL rt_hw_trap_error B .