/* * Copyright (c) 2006-2021, RT-Thread Development Team * * SPDX-License-Identifier: Apache-2.0 * * Change Logs: * Date Author Notes * 2023-10-25 Shell Move ffs to cpuport, add general implementation * by inline assembly * 2024-01-18 Shell support rt_hw_thread_self to improve overall performance */ #ifndef CPUPORT_H__ #define CPUPORT_H__ #include #include #include #ifdef RT_USING_SMP /** * Spinlock */ typedef struct { rt_uint32_t value; } rt_hw_spinlock_t; #endif /* RT_USING_SMP */ #define rt_hw_barrier(cmd, ...) \ __asm__ volatile (RT_STRINGIFY(cmd) " "RT_STRINGIFY(__VA_ARGS__):::"memory") #define rt_hw_isb() rt_hw_barrier(isb) #define rt_hw_dmb() rt_hw_barrier(dmb, ish) #define rt_hw_wmb() rt_hw_barrier(dmb, ishst) #define rt_hw_rmb() rt_hw_barrier(dmb, ishld) #define rt_hw_dsb() rt_hw_barrier(dsb, ish) #define rt_hw_wfi() rt_hw_barrier(wfi) #define rt_hw_wfe() rt_hw_barrier(wfe) #define rt_hw_sev() rt_hw_barrier(sev) #define rt_hw_cpu_relax() rt_hw_barrier(yield) #define rt_hw_sysreg_write(sysreg, val) \ __asm__ volatile ("msr "RT_STRINGIFY(sysreg)", %0"::"r"((rt_uint64_t)(val))) #define rt_hw_sysreg_read(sysreg, val) \ __asm__ volatile ("mrs %0, "RT_STRINGIFY(sysreg)"":"=r"((val))) void _thread_start(void); #ifdef ARCH_USING_HW_THREAD_SELF rt_inline struct rt_thread *rt_hw_thread_self(void) { struct rt_thread *thread; __asm__ volatile ("mrs %0, " RT_STRINGIFY(ARM64_THREAD_REG) :"=r"(thread)); return thread; } rt_inline void rt_hw_thread_set_self(struct rt_thread *thread) { __asm__ volatile ("msr " RT_STRINGIFY(ARM64_THREAD_REG) ", %0"::"r"(thread)); } #endif /* ARCH_USING_HW_THREAD_SELF */ #endif /*CPUPORT_H__*/