/* * Copyright (c) 2006-2021, RT-Thread Development Team * * SPDX-License-Identifier: Apache-2.0 * * Change Logs: * Date Author Notes * 2006-03-18 Bernard the first version * 2006-04-26 Bernard add semaphore APIs * 2006-08-10 Bernard add version information * 2007-01-28 Bernard rename RT_OBJECT_Class_Static to RT_Object_Class_Static * 2007-03-03 Bernard clean up the definitions to rtdef.h * 2010-04-11 yi.qiu add module feature * 2013-06-24 Bernard add rt_kprintf re-define when not use RT_USING_CONSOLE. * 2016-08-09 ArdaFu add new thread and interrupt hook. * 2018-11-22 Jesven add all cpu's lock and ipi handler * 2021-02-28 Meco Man add RT_KSERVICE_USING_STDLIB * 2021-11-14 Meco Man add rtlegacy.h for compatibility * 2022-06-04 Meco Man remove strnlen * 2023-05-20 Bernard add rtatomic.h header file to included files. * 2023-06-30 ChuShicheng move debug check from the rtdebug.h * 2023-10-16 Shell Support a new backtrace framework * 2023-12-10 xqyjlj fix spinlock in up * 2024-01-25 Shell Add rt_susp_list for IPC primitives * 2024-03-10 Meco Man move std libc related functions to rtklibc */ #ifndef __RT_THREAD_H__ #define __RT_THREAD_H__ #include #include #include #include #include #include #ifdef RT_USING_LEGACY #include #endif #ifdef RT_USING_FINSH #include #endif /* RT_USING_FINSH */ #ifdef __cplusplus extern "C" { #endif #ifdef __GNUC__ int entry(void); #endif /** * @addtogroup KernelObject * @{ */ /* * kernel object interface */ struct rt_object_information * rt_object_get_information(enum rt_object_class_type type); int rt_object_get_length(enum rt_object_class_type type); int rt_object_get_pointers(enum rt_object_class_type type, rt_object_t *pointers, int maxlen); void rt_object_init(struct rt_object *object, enum rt_object_class_type type, const char *name); void rt_object_detach(rt_object_t object); #ifdef RT_USING_HEAP rt_object_t rt_object_allocate(enum rt_object_class_type type, const char *name); void rt_object_delete(rt_object_t object); /* custom object */ rt_object_t rt_custom_object_create(const char *name, void *data, rt_err_t (*data_destroy)(void *)); rt_err_t rt_custom_object_destroy(rt_object_t obj); #endif /* RT_USING_HEAP */ rt_bool_t rt_object_is_systemobject(rt_object_t object); rt_uint8_t rt_object_get_type(rt_object_t object); rt_object_t rt_object_find(const char *name, rt_uint8_t type); rt_err_t rt_object_get_name(rt_object_t object, char *name, rt_uint8_t name_size); #ifdef RT_USING_HOOK void rt_object_attach_sethook(void (*hook)(struct rt_object *object)); void rt_object_detach_sethook(void (*hook)(struct rt_object *object)); void rt_object_trytake_sethook(void (*hook)(struct rt_object *object)); void rt_object_take_sethook(void (*hook)(struct rt_object *object)); void rt_object_put_sethook(void (*hook)(struct rt_object *object)); #endif /* RT_USING_HOOK */ /**@}*/ /** * @addtogroup Clock * @{ */ /* * clock & timer interface */ rt_tick_t rt_tick_get(void); void rt_tick_set(rt_tick_t tick); void rt_tick_increase(void); rt_tick_t rt_tick_from_millisecond(rt_int32_t ms); rt_tick_t rt_tick_get_millisecond(void); #ifdef RT_USING_HOOK void rt_tick_sethook(void (*hook)(void)); #endif /* RT_USING_HOOK */ void rt_system_timer_init(void); void rt_system_timer_thread_init(void); void rt_timer_init(rt_timer_t timer, const char *name, void (*timeout)(void *parameter), void *parameter, rt_tick_t time, rt_uint8_t flag); rt_err_t rt_timer_detach(rt_timer_t timer); #ifdef RT_USING_HEAP rt_timer_t rt_timer_create(const char *name, void (*timeout)(void *parameter), void *parameter, rt_tick_t time, rt_uint8_t flag); rt_err_t rt_timer_delete(rt_timer_t timer); #endif /* RT_USING_HEAP */ rt_err_t rt_timer_start(rt_timer_t timer); rt_err_t rt_timer_stop(rt_timer_t timer); rt_err_t rt_timer_control(rt_timer_t timer, int cmd, void *arg); rt_tick_t rt_timer_next_timeout_tick(void); void rt_timer_check(void); #ifdef RT_USING_HOOK void rt_timer_enter_sethook(void (*hook)(struct rt_timer *timer)); void rt_timer_exit_sethook(void (*hook)(struct rt_timer *timer)); #endif /* RT_USING_HOOK */ /**@}*/ /** * @addtogroup Thread * @{ */ /* * thread interface */ rt_err_t rt_thread_init(struct rt_thread *thread, const char *name, void (*entry)(void *parameter), void *parameter, void *stack_start, rt_uint32_t stack_size, rt_uint8_t priority, rt_uint32_t tick); rt_err_t rt_thread_detach(rt_thread_t thread); #ifdef RT_USING_HEAP rt_thread_t rt_thread_create(const char *name, void (*entry)(void *parameter), void *parameter, rt_uint32_t stack_size, rt_uint8_t priority, rt_uint32_t tick); rt_err_t rt_thread_delete(rt_thread_t thread); rt_err_t rt_thread_close(rt_thread_t thread); #endif /* RT_USING_HEAP */ rt_thread_t rt_thread_self(void); rt_thread_t rt_thread_find(char *name); rt_err_t rt_thread_startup(rt_thread_t thread); rt_err_t rt_thread_yield(void); rt_err_t rt_thread_delay(rt_tick_t tick); rt_err_t rt_thread_delay_until(rt_tick_t *tick, rt_tick_t inc_tick); rt_err_t rt_thread_mdelay(rt_int32_t ms); rt_err_t rt_thread_control(rt_thread_t thread, int cmd, void *arg); rt_err_t rt_thread_suspend(rt_thread_t thread); rt_err_t rt_thread_suspend_with_flag(rt_thread_t thread, int suspend_flag); rt_err_t rt_thread_resume(rt_thread_t thread); #ifdef RT_USING_SMART rt_err_t rt_thread_wakeup(rt_thread_t thread); void rt_thread_wakeup_set(struct rt_thread *thread, rt_wakeup_func_t func, void* user_data); #endif /* RT_USING_SMART */ rt_err_t rt_thread_get_name(rt_thread_t thread, char *name, rt_uint8_t name_size); #ifdef RT_USING_SIGNALS void rt_thread_alloc_sig(rt_thread_t tid); void rt_thread_free_sig(rt_thread_t tid); int rt_thread_kill(rt_thread_t tid, int sig); #endif /* RT_USING_SIGNALS */ #ifdef RT_USING_HOOK void rt_thread_suspend_sethook(void (*hook)(rt_thread_t thread)); void rt_thread_resume_sethook (void (*hook)(rt_thread_t thread)); /** * @brief Sets a hook function when a thread is initialized. * * @param thread is the target thread that initializing */ typedef void (*rt_thread_inited_hookproto_t)(rt_thread_t thread); RT_OBJECT_HOOKLIST_DECLARE(rt_thread_inited_hookproto_t, rt_thread_inited); #endif /* RT_USING_HOOK */ /* * idle thread interface */ void rt_thread_idle_init(void); #if defined(RT_USING_HOOK) || defined(RT_USING_IDLE_HOOK) rt_err_t rt_thread_idle_sethook(void (*hook)(void)); rt_err_t rt_thread_idle_delhook(void (*hook)(void)); #endif /* defined(RT_USING_HOOK) || defined(RT_USING_IDLE_HOOK) */ rt_thread_t rt_thread_idle_gethandler(void); /* * schedule service */ void rt_system_scheduler_init(void); void rt_system_scheduler_start(void); void rt_schedule(void); void rt_scheduler_do_irq_switch(void *context); #ifdef RT_USING_OVERFLOW_CHECK void rt_scheduler_stack_check(struct rt_thread *thread); #define RT_SCHEDULER_STACK_CHECK(thr) rt_scheduler_stack_check(thr) #else /* !RT_USING_OVERFLOW_CHECK */ #define RT_SCHEDULER_STACK_CHECK(thr) #endif /* RT_USING_OVERFLOW_CHECK */ rt_base_t rt_enter_critical(void); void rt_exit_critical(void); void rt_exit_critical_safe(rt_base_t critical_level); rt_uint16_t rt_critical_level(void); #ifdef RT_USING_HOOK void rt_scheduler_sethook(void (*hook)(rt_thread_t from, rt_thread_t to)); void rt_scheduler_switch_sethook(void (*hook)(struct rt_thread *tid)); #endif /* RT_USING_HOOK */ #ifdef RT_USING_SMP void rt_secondary_cpu_entry(void); void rt_scheduler_ipi_handler(int vector, void *param); #endif /* RT_USING_SMP */ /**@}*/ /** * @addtogroup Signals * @{ */ #ifdef RT_USING_SIGNALS void rt_signal_mask(int signo); void rt_signal_unmask(int signo); rt_sighandler_t rt_signal_install(int signo, rt_sighandler_t handler); int rt_signal_wait(const rt_sigset_t *set, rt_siginfo_t *si, rt_int32_t timeout); int rt_system_signal_init(void); #endif /* RT_USING_SIGNALS */ /**@}*/ /** * @addtogroup MM * @{ */ /* * memory management interface */ #ifdef RT_USING_MEMPOOL /* * memory pool interface */ rt_err_t rt_mp_init(struct rt_mempool *mp, const char *name, void *start, rt_size_t size, rt_size_t block_size); rt_err_t rt_mp_detach(struct rt_mempool *mp); #ifdef RT_USING_HEAP rt_mp_t rt_mp_create(const char *name, rt_size_t block_count, rt_size_t block_size); rt_err_t rt_mp_delete(rt_mp_t mp); #endif /* RT_USING_HEAP */ void *rt_mp_alloc(rt_mp_t mp, rt_int32_t time); void rt_mp_free(void *block); #ifdef RT_USING_HOOK void rt_mp_alloc_sethook(void (*hook)(struct rt_mempool *mp, void *block)); void rt_mp_free_sethook(void (*hook)(struct rt_mempool *mp, void *block)); #endif /* RT_USING_HOOK */ #endif /* RT_USING_MEMPOOL */ #ifdef RT_USING_HEAP /* * heap memory interface */ void rt_system_heap_init(void *begin_addr, void *end_addr); void *rt_malloc(rt_size_t size); void rt_free(void *ptr); void *rt_realloc(void *ptr, rt_size_t newsize); void *rt_calloc(rt_size_t count, rt_size_t size); void *rt_malloc_align(rt_size_t size, rt_size_t align); void rt_free_align(void *ptr); void rt_memory_info(rt_size_t *total, rt_size_t *used, rt_size_t *max_used); #if defined(RT_USING_SLAB) && defined(RT_USING_SLAB_AS_HEAP) void *rt_page_alloc(rt_size_t npages); void rt_page_free(void *addr, rt_size_t npages); #endif /* defined(RT_USING_SLAB) && defined(RT_USING_SLAB_AS_HEAP) */ #ifdef RT_USING_HOOK void rt_malloc_sethook(void (*hook)(void **ptr, rt_size_t size)); void rt_realloc_set_entry_hook(void (*hook)(void **ptr, rt_size_t size)); void rt_realloc_set_exit_hook(void (*hook)(void **ptr, rt_size_t size)); void rt_free_sethook(void (*hook)(void **ptr)); #endif /* RT_USING_HOOK */ #endif /* RT_USING_HEAP */ #ifdef RT_USING_SMALL_MEM /** * small memory object interface */ rt_smem_t rt_smem_init(const char *name, void *begin_addr, rt_size_t size); rt_err_t rt_smem_detach(rt_smem_t m); void *rt_smem_alloc(rt_smem_t m, rt_size_t size); void *rt_smem_realloc(rt_smem_t m, void *rmem, rt_size_t newsize); void rt_smem_free(void *rmem); #endif /* RT_USING_SMALL_MEM */ #ifdef RT_USING_MEMHEAP /** * memory heap object interface */ rt_err_t rt_memheap_init(struct rt_memheap *memheap, const char *name, void *start_addr, rt_size_t size); rt_err_t rt_memheap_detach(struct rt_memheap *heap); void *rt_memheap_alloc(struct rt_memheap *heap, rt_size_t size); void *rt_memheap_realloc(struct rt_memheap *heap, void *ptr, rt_size_t newsize); void rt_memheap_free(void *ptr); void rt_memheap_info(struct rt_memheap *heap, rt_size_t *total, rt_size_t *used, rt_size_t *max_used); #endif /* RT_USING_MEMHEAP */ #ifdef RT_USING_MEMHEAP_AS_HEAP /** * memory heap as heap */ void *_memheap_alloc(struct rt_memheap *heap, rt_size_t size); void _memheap_free(void *rmem); void *_memheap_realloc(struct rt_memheap *heap, void *rmem, rt_size_t newsize); #endif #ifdef RT_USING_SLAB /** * slab object interface */ rt_slab_t rt_slab_init(const char *name, void *begin_addr, rt_size_t size); rt_err_t rt_slab_detach(rt_slab_t m); void *rt_slab_page_alloc(rt_slab_t m, rt_size_t npages); void rt_slab_page_free(rt_slab_t m, void *addr, rt_size_t npages); void *rt_slab_alloc(rt_slab_t m, rt_size_t size); void *rt_slab_realloc(rt_slab_t m, void *ptr, rt_size_t size); void rt_slab_free(rt_slab_t m, void *ptr); #endif /* RT_USING_SLAB */ /**@}*/ /** * @addtogroup IPC * @{ */ /** * Suspend list - A basic building block for IPC primitives which interacts with * scheduler directly. Its API is similar to a FIFO list. * * Note: don't use in application codes directly */ void rt_susp_list_print(rt_list_t *list); /* reserve thread error while resuming it */ #define RT_THREAD_RESUME_RES_THR_ERR (-1) struct rt_thread *rt_susp_list_dequeue(rt_list_t *susp_list, rt_err_t thread_error); rt_err_t rt_susp_list_resume_all(rt_list_t *susp_list, rt_err_t thread_error); rt_err_t rt_susp_list_resume_all_irq(rt_list_t *susp_list, rt_err_t thread_error, struct rt_spinlock *lock); /* suspend and enqueue */ rt_err_t rt_thread_suspend_to_list(rt_thread_t thread, rt_list_t *susp_list, int ipc_flags, int suspend_flag); /* only for a suspended thread, and caller must hold the scheduler lock */ rt_err_t rt_susp_list_enqueue(rt_list_t *susp_list, rt_thread_t thread, int ipc_flags); #ifdef RT_USING_SEMAPHORE /* * semaphore interface */ rt_err_t rt_sem_init(rt_sem_t sem, const char *name, rt_uint32_t value, rt_uint8_t flag); rt_err_t rt_sem_detach(rt_sem_t sem); #ifdef RT_USING_HEAP rt_sem_t rt_sem_create(const char *name, rt_uint32_t value, rt_uint8_t flag); rt_err_t rt_sem_delete(rt_sem_t sem); #endif /* RT_USING_HEAP */ rt_err_t rt_sem_take(rt_sem_t sem, rt_int32_t timeout); rt_err_t rt_sem_take_interruptible(rt_sem_t sem, rt_int32_t timeout); rt_err_t rt_sem_take_killable(rt_sem_t sem, rt_int32_t timeout); rt_err_t rt_sem_trytake(rt_sem_t sem); rt_err_t rt_sem_release(rt_sem_t sem); rt_err_t rt_sem_control(rt_sem_t sem, int cmd, void *arg); #endif /* RT_USING_SEMAPHORE */ #ifdef RT_USING_MUTEX /* * mutex interface */ rt_err_t rt_mutex_init(rt_mutex_t mutex, const char *name, rt_uint8_t flag); rt_err_t rt_mutex_detach(rt_mutex_t mutex); #ifdef RT_USING_HEAP rt_mutex_t rt_mutex_create(const char *name, rt_uint8_t flag); rt_err_t rt_mutex_delete(rt_mutex_t mutex); #endif /* RT_USING_HEAP */ void rt_mutex_drop_thread(rt_mutex_t mutex, rt_thread_t thread); rt_uint8_t rt_mutex_setprioceiling(rt_mutex_t mutex, rt_uint8_t priority); rt_uint8_t rt_mutex_getprioceiling(rt_mutex_t mutex); rt_err_t rt_mutex_take(rt_mutex_t mutex, rt_int32_t timeout); rt_err_t rt_mutex_trytake(rt_mutex_t mutex); rt_err_t rt_mutex_take_interruptible(rt_mutex_t mutex, rt_int32_t time); rt_err_t rt_mutex_take_killable(rt_mutex_t mutex, rt_int32_t time); rt_err_t rt_mutex_release(rt_mutex_t mutex); rt_err_t rt_mutex_control(rt_mutex_t mutex, int cmd, void *arg); rt_inline rt_thread_t rt_mutex_get_owner(rt_mutex_t mutex) { return mutex->owner; } rt_inline rt_ubase_t rt_mutex_get_hold(rt_mutex_t mutex) { return mutex->hold; } #endif /* RT_USING_MUTEX */ #ifdef RT_USING_EVENT /* * event interface */ rt_err_t rt_event_init(rt_event_t event, const char *name, rt_uint8_t flag); rt_err_t rt_event_detach(rt_event_t event); #ifdef RT_USING_HEAP rt_event_t rt_event_create(const char *name, rt_uint8_t flag); rt_err_t rt_event_delete(rt_event_t event); #endif /* RT_USING_HEAP */ rt_err_t rt_event_send(rt_event_t event, rt_uint32_t set); rt_err_t rt_event_recv(rt_event_t event, rt_uint32_t set, rt_uint8_t opt, rt_int32_t timeout, rt_uint32_t *recved); rt_err_t rt_event_recv_interruptible(rt_event_t event, rt_uint32_t set, rt_uint8_t opt, rt_int32_t timeout, rt_uint32_t *recved); rt_err_t rt_event_recv_killable(rt_event_t event, rt_uint32_t set, rt_uint8_t opt, rt_int32_t timeout, rt_uint32_t *recved); rt_err_t rt_event_control(rt_event_t event, int cmd, void *arg); #endif /* RT_USING_EVENT */ #ifdef RT_USING_MAILBOX /* * mailbox interface */ rt_err_t rt_mb_init(rt_mailbox_t mb, const char *name, void *msgpool, rt_size_t size, rt_uint8_t flag); rt_err_t rt_mb_detach(rt_mailbox_t mb); #ifdef RT_USING_HEAP rt_mailbox_t rt_mb_create(const char *name, rt_size_t size, rt_uint8_t flag); rt_err_t rt_mb_delete(rt_mailbox_t mb); #endif /* RT_USING_HEAP */ rt_err_t rt_mb_send(rt_mailbox_t mb, rt_ubase_t value); rt_err_t rt_mb_send_interruptible(rt_mailbox_t mb, rt_ubase_t value); rt_err_t rt_mb_send_killable(rt_mailbox_t mb, rt_ubase_t value); rt_err_t rt_mb_send_wait(rt_mailbox_t mb, rt_ubase_t value, rt_int32_t timeout); rt_err_t rt_mb_send_wait_interruptible(rt_mailbox_t mb, rt_ubase_t value, rt_int32_t timeout); rt_err_t rt_mb_send_wait_killable(rt_mailbox_t mb, rt_ubase_t value, rt_int32_t timeout); rt_err_t rt_mb_urgent(rt_mailbox_t mb, rt_ubase_t value); rt_err_t rt_mb_recv(rt_mailbox_t mb, rt_ubase_t *value, rt_int32_t timeout); rt_err_t rt_mb_recv_interruptible(rt_mailbox_t mb, rt_ubase_t *value, rt_int32_t timeout); rt_err_t rt_mb_recv_killable(rt_mailbox_t mb, rt_ubase_t *value, rt_int32_t timeout); rt_err_t rt_mb_control(rt_mailbox_t mb, int cmd, void *arg); #endif /* RT_USING_MAILBOX */ #ifdef RT_USING_MESSAGEQUEUE struct rt_mq_message { struct rt_mq_message *next; rt_ssize_t length; #ifdef RT_USING_MESSAGEQUEUE_PRIORITY rt_int32_t prio; #endif /* RT_USING_MESSAGEQUEUE_PRIORITY */ }; #define RT_MQ_BUF_SIZE(msg_size, max_msgs) \ ((RT_ALIGN((msg_size), RT_ALIGN_SIZE) + sizeof(struct rt_mq_message)) * (max_msgs)) /* * message queue interface */ rt_err_t rt_mq_init(rt_mq_t mq, const char *name, void *msgpool, rt_size_t msg_size, rt_size_t pool_size, rt_uint8_t flag); rt_err_t rt_mq_detach(rt_mq_t mq); #ifdef RT_USING_HEAP rt_mq_t rt_mq_create(const char *name, rt_size_t msg_size, rt_size_t max_msgs, rt_uint8_t flag); rt_err_t rt_mq_delete(rt_mq_t mq); #endif /* RT_USING_HEAP */ rt_err_t rt_mq_send(rt_mq_t mq, const void *buffer, rt_size_t size); rt_err_t rt_mq_send_interruptible(rt_mq_t mq, const void *buffer, rt_size_t size); rt_err_t rt_mq_send_killable(rt_mq_t mq, const void *buffer, rt_size_t size); rt_err_t rt_mq_send_wait(rt_mq_t mq, const void *buffer, rt_size_t size, rt_int32_t timeout); rt_err_t rt_mq_send_wait_interruptible(rt_mq_t mq, const void *buffer, rt_size_t size, rt_int32_t timeout); rt_err_t rt_mq_send_wait_killable(rt_mq_t mq, const void *buffer, rt_size_t size, rt_int32_t timeout); rt_err_t rt_mq_urgent(rt_mq_t mq, const void *buffer, rt_size_t size); rt_ssize_t rt_mq_recv(rt_mq_t mq, void *buffer, rt_size_t size, rt_int32_t timeout); rt_ssize_t rt_mq_recv_interruptible(rt_mq_t mq, void *buffer, rt_size_t size, rt_int32_t timeout); rt_ssize_t rt_mq_recv_killable(rt_mq_t mq, void *buffer, rt_size_t size, rt_int32_t timeout); rt_err_t rt_mq_control(rt_mq_t mq, int cmd, void *arg); #ifdef RT_USING_MESSAGEQUEUE_PRIORITY rt_err_t rt_mq_send_wait_prio(rt_mq_t mq, const void *buffer, rt_size_t size, rt_int32_t prio, rt_int32_t timeout, int suspend_flag); rt_ssize_t rt_mq_recv_prio(rt_mq_t mq, void *buffer, rt_size_t size, rt_int32_t *prio, rt_int32_t timeout, int suspend_flag); #endif /* RT_USING_MESSAGEQUEUE_PRIORITY */ #endif /* RT_USING_MESSAGEQUEUE */ /* defunct */ void rt_thread_defunct_enqueue(rt_thread_t thread); rt_thread_t rt_thread_defunct_dequeue(void); /* * spinlock */ struct rt_spinlock; void rt_spin_lock_init(struct rt_spinlock *lock); void rt_spin_lock(struct rt_spinlock *lock); void rt_spin_unlock(struct rt_spinlock *lock); rt_base_t rt_spin_lock_irqsave(struct rt_spinlock *lock); void rt_spin_unlock_irqrestore(struct rt_spinlock *lock, rt_base_t level); /**@}*/ #ifdef RT_USING_DEVICE /** * @addtogroup Device * @{ */ /* * device (I/O) system interface */ rt_device_t rt_device_find(const char *name); rt_err_t rt_device_register(rt_device_t dev, const char *name, rt_uint16_t flags); rt_err_t rt_device_unregister(rt_device_t dev); #ifdef RT_USING_HEAP rt_device_t rt_device_create(int type, int attach_size); void rt_device_destroy(rt_device_t device); #endif /* RT_USING_HEAP */ rt_err_t rt_device_set_rx_indicate(rt_device_t dev, rt_err_t (*rx_ind)(rt_device_t dev, rt_size_t size)); rt_err_t rt_device_set_tx_complete(rt_device_t dev, rt_err_t (*tx_done)(rt_device_t dev, void *buffer)); rt_err_t rt_device_init (rt_device_t dev); rt_err_t rt_device_open (rt_device_t dev, rt_uint16_t oflag); rt_err_t rt_device_close(rt_device_t dev); rt_ssize_t rt_device_read(rt_device_t dev, rt_off_t pos, void *buffer, rt_size_t size); rt_ssize_t rt_device_write(rt_device_t dev, rt_off_t pos, const void *buffer, rt_size_t size); rt_err_t rt_device_control(rt_device_t dev, int cmd, void *arg); /**@}*/ #endif /* RT_USING_DEVICE */ /* * interrupt service */ /* * rt_interrupt_enter and rt_interrupt_leave only can be called by BSP */ void rt_interrupt_enter(void); void rt_interrupt_leave(void); #ifdef RT_USING_SMP /* * smp cpus lock service */ rt_base_t rt_cpus_lock(void); void rt_cpus_unlock(rt_base_t level); void rt_cpus_lock_status_restore(struct rt_thread *thread); struct rt_cpu *rt_cpu_self(void); struct rt_cpu *rt_cpu_index(int index); #ifdef RT_USING_DEBUG rt_base_t rt_cpu_get_id(void); #else /* !RT_USING_DEBUG */ #define rt_cpu_get_id rt_hw_cpu_id #endif /* RT_USING_DEBUG */ #else /* !RT_USING_SMP */ #define rt_cpu_get_id() (0) #endif /* RT_USING_SMP */ /* * the number of nested interrupts. */ rt_uint8_t rt_interrupt_get_nest(void); #ifdef RT_USING_HOOK void rt_interrupt_enter_sethook(void (*hook)(void)); void rt_interrupt_leave_sethook(void (*hook)(void)); #endif /* RT_USING_HOOK */ #ifdef RT_USING_COMPONENTS_INIT void rt_components_init(void); void rt_components_board_init(void); #endif /* RT_USING_COMPONENTS_INIT */ /** * @addtogroup KernelService * @{ */ /* * general kernel service */ #ifndef RT_USING_CONSOLE #define rt_kprintf(...) #define rt_kputs(str) #else int rt_kprintf(const char *fmt, ...); void rt_kputs(const char *str); #endif /* RT_USING_CONSOLE */ rt_err_t rt_backtrace(void); rt_err_t rt_backtrace_thread(rt_thread_t thread); rt_err_t rt_backtrace_frame(struct rt_hw_backtrace_frame *frame); #if defined(RT_USING_DEVICE) && defined(RT_USING_CONSOLE) rt_device_t rt_console_set_device(const char *name); rt_device_t rt_console_get_device(void); #ifdef RT_USING_THREADSAFE_PRINTF rt_thread_t rt_console_current_user(void); #else rt_inline void *rt_console_current_user(void) { return RT_NULL; } #endif /* RT_USING_THREADSAFE_PRINTF */ #endif /* defined(RT_USING_DEVICE) && defined(RT_USING_CONSOLE) */ int __rt_ffs(int value); void rt_show_version(void); #ifdef RT_DEBUGING_ASSERT extern void (*rt_assert_hook)(const char *ex, const char *func, rt_size_t line); void rt_assert_set_hook(void (*hook)(const char *ex, const char *func, rt_size_t line)); void rt_assert_handler(const char *ex, const char *func, rt_size_t line); #define RT_ASSERT(EX) \ if (!(EX)) \ { \ rt_assert_handler(#EX, __FUNCTION__, __LINE__); \ } #else #define RT_ASSERT(EX) #endif /* RT_DEBUGING_ASSERT */ #ifdef RT_DEBUGING_CONTEXT /* Macro to check current context */ #define RT_DEBUG_NOT_IN_INTERRUPT \ do \ { \ if (rt_interrupt_get_nest() != 0) \ { \ rt_kprintf("Function[%s] shall not be used in ISR\n", __FUNCTION__); \ RT_ASSERT(0) \ } \ } \ while (0) /* "In thread context" means: * 1) the scheduler has been started * 2) not in interrupt context. */ #define RT_DEBUG_IN_THREAD_CONTEXT \ do \ { \ if (rt_thread_self() == RT_NULL) \ { \ rt_kprintf("Function[%s] shall not be used before scheduler start\n", \ __FUNCTION__); \ RT_ASSERT(0) \ } \ RT_DEBUG_NOT_IN_INTERRUPT; \ } \ while (0) /* "scheduler available" means: * 1) the scheduler has been started. * 2) not in interrupt context. * 3) scheduler is not locked. */ #define RT_DEBUG_SCHEDULER_AVAILABLE(need_check) \ do \ { \ if (need_check) \ { \ if (rt_critical_level() != 0) \ { \ rt_kprintf("Function[%s]: scheduler is not available\n", \ __FUNCTION__); \ RT_ASSERT(0) \ } \ RT_DEBUG_IN_THREAD_CONTEXT; \ } \ } \ while (0) #else #define RT_DEBUG_NOT_IN_INTERRUPT #define RT_DEBUG_IN_THREAD_CONTEXT #define RT_DEBUG_SCHEDULER_AVAILABLE(need_check) #endif /* RT_DEBUGING_CONTEXT */ rt_inline rt_bool_t rt_in_thread_context(void) { return rt_thread_self() != RT_NULL && rt_interrupt_get_nest() == 0; } /* is scheduler available */ rt_inline rt_bool_t rt_scheduler_is_available(void) { return rt_critical_level() == 0 && rt_in_thread_context(); } #ifdef RT_USING_SMP /* is thread bond on core */ rt_inline rt_bool_t rt_sched_thread_is_binding(rt_thread_t thread) { if (thread == RT_NULL) { thread = rt_thread_self(); } return !thread || RT_SCHED_CTX(thread).bind_cpu != RT_CPUS_NR; } #else #define rt_sched_thread_is_binding(thread) (RT_TRUE) #endif /**@}*/ #ifdef __cplusplus } #endif #endif /* __RT_THREAD_H__ */