252 lines
6.4 KiB
C
252 lines
6.4 KiB
C
|
/*
|
||
|
* Copyright (c) 2006-2021, RT-Thread Development Team
|
||
|
*
|
||
|
* SPDX-License-Identifier: Apache-2.0
|
||
|
*
|
||
|
* Change Logs:
|
||
|
* Date Author Notes
|
||
|
* 2006-03-18 Bernard the first version
|
||
|
* 2006-04-25 Bernard add rt_hw_context_switch_interrupt declaration
|
||
|
* 2006-09-24 Bernard add rt_hw_context_switch_to declaration
|
||
|
* 2012-12-29 Bernard add rt_hw_exception_install declaration
|
||
|
* 2017-10-17 Hichard add some macros
|
||
|
* 2018-11-17 Jesven add rt_hw_spinlock_t
|
||
|
* add smp support
|
||
|
* 2019-05-18 Bernard add empty definition for not enable cache case
|
||
|
* 2023-09-15 xqyjlj perf rt_hw_interrupt_disable/enable
|
||
|
* 2023-10-16 Shell Support a new backtrace framework
|
||
|
*/
|
||
|
|
||
|
#ifndef __RT_HW_H__
|
||
|
#define __RT_HW_H__
|
||
|
|
||
|
#include <rtdef.h>
|
||
|
|
||
|
#if defined (RT_USING_CACHE) || defined(RT_USING_SMP) || defined(RT_HW_INCLUDE_CPUPORT)
|
||
|
#include <cpuport.h> /* include spinlock, cache ops, etc. */
|
||
|
#endif
|
||
|
|
||
|
#ifdef __cplusplus
|
||
|
extern "C" {
|
||
|
#endif
|
||
|
|
||
|
/*
|
||
|
* Some macros define
|
||
|
*/
|
||
|
#ifndef HWREG64
|
||
|
#define HWREG64(x) (*((volatile rt_uint64_t *)(x)))
|
||
|
#endif
|
||
|
#ifndef HWREG32
|
||
|
#define HWREG32(x) (*((volatile rt_uint32_t *)(x)))
|
||
|
#endif
|
||
|
#ifndef HWREG16
|
||
|
#define HWREG16(x) (*((volatile rt_uint16_t *)(x)))
|
||
|
#endif
|
||
|
#ifndef HWREG8
|
||
|
#define HWREG8(x) (*((volatile rt_uint8_t *)(x)))
|
||
|
#endif
|
||
|
|
||
|
#ifndef RT_CPU_CACHE_LINE_SZ
|
||
|
#define RT_CPU_CACHE_LINE_SZ 32
|
||
|
#endif
|
||
|
|
||
|
enum RT_HW_CACHE_OPS
|
||
|
{
|
||
|
RT_HW_CACHE_FLUSH = 0x01,
|
||
|
RT_HW_CACHE_INVALIDATE = 0x02,
|
||
|
};
|
||
|
|
||
|
/*
|
||
|
* CPU interfaces
|
||
|
*/
|
||
|
#ifdef RT_USING_CACHE
|
||
|
|
||
|
#ifdef RT_USING_SMART
|
||
|
#include <cache.h>
|
||
|
#endif
|
||
|
|
||
|
void rt_hw_cpu_icache_enable(void);
|
||
|
void rt_hw_cpu_icache_disable(void);
|
||
|
rt_base_t rt_hw_cpu_icache_status(void);
|
||
|
void rt_hw_cpu_icache_ops(int ops, void* addr, int size);
|
||
|
|
||
|
void rt_hw_cpu_dcache_enable(void);
|
||
|
void rt_hw_cpu_dcache_disable(void);
|
||
|
rt_base_t rt_hw_cpu_dcache_status(void);
|
||
|
void rt_hw_cpu_dcache_ops(int ops, void* addr, int size);
|
||
|
#else
|
||
|
|
||
|
/* define cache ops as empty */
|
||
|
#define rt_hw_cpu_icache_enable(...)
|
||
|
#define rt_hw_cpu_icache_disable(...)
|
||
|
#define rt_hw_cpu_icache_ops(...)
|
||
|
#define rt_hw_cpu_dcache_enable(...)
|
||
|
#define rt_hw_cpu_dcache_disable(...)
|
||
|
#define rt_hw_cpu_dcache_ops(...)
|
||
|
|
||
|
#define rt_hw_cpu_icache_status(...) 0
|
||
|
#define rt_hw_cpu_dcache_status(...) 0
|
||
|
|
||
|
#endif
|
||
|
|
||
|
void rt_hw_cpu_reset(void);
|
||
|
void rt_hw_cpu_shutdown(void);
|
||
|
|
||
|
const char *rt_hw_cpu_arch(void);
|
||
|
|
||
|
rt_uint8_t *rt_hw_stack_init(void *entry,
|
||
|
void *parameter,
|
||
|
rt_uint8_t *stack_addr,
|
||
|
void *exit);
|
||
|
|
||
|
#ifdef RT_USING_HW_STACK_GUARD
|
||
|
void rt_hw_stack_guard_init(rt_thread_t thread);
|
||
|
#endif
|
||
|
|
||
|
/*
|
||
|
* Interrupt handler definition
|
||
|
*/
|
||
|
typedef void (*rt_isr_handler_t)(int vector, void *param);
|
||
|
|
||
|
struct rt_irq_desc
|
||
|
{
|
||
|
rt_isr_handler_t handler;
|
||
|
void *param;
|
||
|
|
||
|
#ifdef RT_USING_INTERRUPT_INFO
|
||
|
char name[RT_NAME_MAX];
|
||
|
rt_uint32_t counter;
|
||
|
#ifdef RT_USING_SMP
|
||
|
rt_ubase_t cpu_counter[RT_CPUS_NR];
|
||
|
#endif
|
||
|
#endif
|
||
|
};
|
||
|
|
||
|
/*
|
||
|
* Interrupt interfaces
|
||
|
*/
|
||
|
void rt_hw_interrupt_init(void);
|
||
|
void rt_hw_interrupt_mask(int vector);
|
||
|
void rt_hw_interrupt_umask(int vector);
|
||
|
rt_isr_handler_t rt_hw_interrupt_install(int vector,
|
||
|
rt_isr_handler_t handler,
|
||
|
void *param,
|
||
|
const char *name);
|
||
|
void rt_hw_interrupt_uninstall(int vector,
|
||
|
rt_isr_handler_t handler,
|
||
|
void *param);
|
||
|
|
||
|
#ifdef RT_USING_SMP
|
||
|
rt_base_t rt_hw_local_irq_disable(void);
|
||
|
void rt_hw_local_irq_enable(rt_base_t level);
|
||
|
|
||
|
rt_base_t rt_cpus_lock(void);
|
||
|
void rt_cpus_unlock(rt_base_t level);
|
||
|
|
||
|
#define rt_hw_interrupt_disable rt_cpus_lock
|
||
|
#define rt_hw_interrupt_enable rt_cpus_unlock
|
||
|
#else
|
||
|
rt_base_t rt_hw_interrupt_disable(void);
|
||
|
void rt_hw_interrupt_enable(rt_base_t level);
|
||
|
|
||
|
#define rt_hw_local_irq_disable rt_hw_interrupt_disable
|
||
|
#define rt_hw_local_irq_enable rt_hw_interrupt_enable
|
||
|
|
||
|
#endif /*RT_USING_SMP*/
|
||
|
rt_bool_t rt_hw_interrupt_is_disabled(void);
|
||
|
|
||
|
/*
|
||
|
* Context interfaces
|
||
|
*/
|
||
|
#ifdef RT_USING_SMP
|
||
|
void rt_hw_context_switch(rt_ubase_t from, rt_ubase_t to, struct rt_thread *to_thread);
|
||
|
void rt_hw_context_switch_to(rt_ubase_t to, struct rt_thread *to_thread);
|
||
|
void rt_hw_context_switch_interrupt(void *context, rt_ubase_t from, rt_ubase_t to, struct rt_thread *to_thread);
|
||
|
#else
|
||
|
void rt_hw_context_switch(rt_ubase_t from, rt_ubase_t to);
|
||
|
void rt_hw_context_switch_to(rt_ubase_t to);
|
||
|
void rt_hw_context_switch_interrupt(rt_ubase_t from, rt_ubase_t to, rt_thread_t from_thread, rt_thread_t to_thread);
|
||
|
#endif /*RT_USING_SMP*/
|
||
|
|
||
|
/**
|
||
|
* Hardware Layer Backtrace Service
|
||
|
*/
|
||
|
struct rt_hw_backtrace_frame {
|
||
|
rt_base_t fp;
|
||
|
rt_base_t pc;
|
||
|
};
|
||
|
|
||
|
rt_err_t rt_hw_backtrace_frame_get(rt_thread_t thread, struct rt_hw_backtrace_frame *frame);
|
||
|
|
||
|
rt_err_t rt_hw_backtrace_frame_unwind(rt_thread_t thread, struct rt_hw_backtrace_frame *frame);
|
||
|
|
||
|
void rt_hw_console_output(const char *str);
|
||
|
|
||
|
void rt_hw_show_memory(rt_uint32_t addr, rt_size_t size);
|
||
|
|
||
|
/*
|
||
|
* Exception interfaces
|
||
|
*/
|
||
|
void rt_hw_exception_install(rt_err_t (*exception_handle)(void *context));
|
||
|
|
||
|
/*
|
||
|
* delay interfaces
|
||
|
*/
|
||
|
void rt_hw_us_delay(rt_uint32_t us);
|
||
|
|
||
|
int rt_hw_cpu_id(void);
|
||
|
|
||
|
#if defined(RT_USING_SMP) || defined(RT_USING_AMP)
|
||
|
/**
|
||
|
* ipi function
|
||
|
*/
|
||
|
void rt_hw_ipi_send(int ipi_vector, unsigned int cpu_mask);
|
||
|
#endif
|
||
|
|
||
|
#ifdef RT_USING_SMP
|
||
|
|
||
|
void rt_hw_spin_lock_init(rt_hw_spinlock_t *lock);
|
||
|
void rt_hw_spin_lock(rt_hw_spinlock_t *lock);
|
||
|
void rt_hw_spin_unlock(rt_hw_spinlock_t *lock);
|
||
|
|
||
|
extern rt_hw_spinlock_t _cpus_lock;
|
||
|
|
||
|
#define __RT_HW_SPIN_LOCK_INITIALIZER(lockname) {0}
|
||
|
|
||
|
#define __RT_HW_SPIN_LOCK_UNLOCKED(lockname) \
|
||
|
(rt_hw_spinlock_t) __RT_HW_SPIN_LOCK_INITIALIZER(lockname)
|
||
|
|
||
|
#define RT_DEFINE_HW_SPINLOCK(x) rt_hw_spinlock_t x = __RT_HW_SPIN_LOCK_UNLOCKED(x)
|
||
|
|
||
|
/**
|
||
|
* boot secondary cpu
|
||
|
*/
|
||
|
void rt_hw_secondary_cpu_up(void);
|
||
|
|
||
|
/**
|
||
|
* secondary cpu idle function
|
||
|
*/
|
||
|
void rt_hw_secondary_cpu_idle_exec(void);
|
||
|
|
||
|
#else /* !RT_USING_SMP */
|
||
|
|
||
|
#define RT_DEFINE_HW_SPINLOCK(x) rt_ubase_t x
|
||
|
|
||
|
#define rt_hw_spin_lock(lock) *(lock) = rt_hw_interrupt_disable()
|
||
|
#define rt_hw_spin_unlock(lock) rt_hw_interrupt_enable(*(lock))
|
||
|
|
||
|
|
||
|
#endif /* RT_USING_SMP */
|
||
|
|
||
|
#ifndef RT_USING_CACHE
|
||
|
#define rt_hw_isb()
|
||
|
#define rt_hw_dmb()
|
||
|
#define rt_hw_dsb()
|
||
|
#endif /* RT_USING_CACHE */
|
||
|
|
||
|
#ifdef __cplusplus
|
||
|
}
|
||
|
#endif
|
||
|
|
||
|
#endif
|