/* * Copyright (c) 2006-2020, RT-Thread Development Team * * SPDX-License-Identifier: Apache-2.0 * * Change Logs: * Date Author Notes * 2024-03-28 Shell Move vector handling codes from context_gcc.S * 2024-04-08 Shell Optimizing exception switch between u-space/kernel, */ #ifndef __ARM64_INC_VECTOR_H__ #define __ARM64_INC_VECTOR_H__ #include "asm-generic.h" #include #include #include .macro SAVE_IRQ_CONTEXT /* Save the entire context. */ SAVE_FPU sp stp x0, x1, [sp, #-0x10]! stp x2, x3, [sp, #-0x10]! stp x4, x5, [sp, #-0x10]! stp x6, x7, [sp, #-0x10]! stp x8, x9, [sp, #-0x10]! stp x10, x11, [sp, #-0x10]! stp x12, x13, [sp, #-0x10]! stp x14, x15, [sp, #-0x10]! stp x16, x17, [sp, #-0x10]! stp x18, x19, [sp, #-0x10]! stp x20, x21, [sp, #-0x10]! stp x22, x23, [sp, #-0x10]! stp x24, x25, [sp, #-0x10]! stp x26, x27, [sp, #-0x10]! stp x28, x29, [sp, #-0x10]! mrs x28, fpcr mrs x29, fpsr stp x28, x29, [sp, #-0x10]! mrs x29, sp_el0 stp x29, x30, [sp, #-0x10]! mrs x3, spsr_el1 mrs x2, elr_el1 stp x2, x3, [sp, #-0x10]! .endm #ifdef RT_USING_SMP #include "../mp/context_gcc.h" #else #include "../up/context_gcc.h" #endif .macro RESTORE_IRQ_CONTEXT_NO_SPEL0 ldp x2, x3, [sp], #0x10 msr elr_el1, x2 msr spsr_el1, x3 ldp x29, x30, [sp], #0x10 ldp x28, x29, [sp], #0x10 msr fpcr, x28 msr fpsr, x29 ldp x28, x29, [sp], #0x10 ldp x26, x27, [sp], #0x10 ldp x24, x25, [sp], #0x10 ldp x22, x23, [sp], #0x10 ldp x20, x21, [sp], #0x10 ldp x18, x19, [sp], #0x10 ldp x16, x17, [sp], #0x10 ldp x14, x15, [sp], #0x10 ldp x12, x13, [sp], #0x10 ldp x10, x11, [sp], #0x10 ldp x8, x9, [sp], #0x10 ldp x6, x7, [sp], #0x10 ldp x4, x5, [sp], #0x10 ldp x2, x3, [sp], #0x10 ldp x0, x1, [sp], #0x10 RESTORE_FPU sp .endm .macro EXCEPTION_SWITCH, eframex, tmpx #ifdef RT_USING_SMART /** * test the spsr for execution level 0 * That is { PSTATE.[NZCV] := SPSR_EL1 & M.EL0t } */ ldr \tmpx, [\eframex, #CONTEXT_OFFSET_SPSR_EL1] and \tmpx, \tmpx, 0x1f cbz \tmpx, 1f b 2f 1: b arch_ret_to_user 2: #endif /* RT_USING_SMART */ .endm .macro SAVE_USER_CTX, eframex, tmpx #ifdef RT_USING_SMART mrs \tmpx, spsr_el1 and \tmpx, \tmpx, 0xf cbz \tmpx, 1f b 2f 1: mov x0, \eframex bl lwp_uthread_ctx_save 2: #endif /* RT_USING_SMART */ .endm .macro RESTORE_USER_CTX, eframex, tmpx #ifdef RT_USING_SMART ldr \tmpx, [\eframex, #CONTEXT_OFFSET_SPSR_EL1] and \tmpx, \tmpx, 0x1f cbz \tmpx, 1f b 2f 1: bl lwp_uthread_ctx_restore 2: #endif /* RT_USING_SMART */ .endm #endif /* __ARM64_INC_VECTOR_H__ */