351 lines
9.3 KiB
C
351 lines
9.3 KiB
C
/*
|
|
* Copyright (c) 2006-2021, RT-Thread Development Team
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*
|
|
* Change Logs:
|
|
* Date Author Notes
|
|
* 2020-11-18 Jesven first version
|
|
* 2021-02-03 lizhirui port to riscv64
|
|
* 2021-02-06 lizhirui add thread filter
|
|
* 2021-02-19 lizhirui port to new version of rt-smart
|
|
* 2021-03-02 lizhirui add a auxillary function for interrupt
|
|
* 2021-03-04 lizhirui delete thread filter
|
|
* 2021-03-04 lizhirui modify for new version of rt-smart
|
|
* 2021-11-22 JasonHu add lwp_set_thread_context
|
|
* 2021-11-30 JasonHu add clone/fork support
|
|
* 2023-07-16 Shell Move part of the codes to C from asm in signal handling
|
|
* 2023-10-16 Shell Support a new backtrace framework
|
|
*/
|
|
#include <rthw.h>
|
|
#include <rtthread.h>
|
|
|
|
#include <stddef.h>
|
|
|
|
#ifdef ARCH_MM_MMU
|
|
|
|
#define DBG_TAG "lwp.arch"
|
|
#define DBG_LVL DBG_INFO
|
|
#include <rtdbg.h>
|
|
|
|
#include <lwp_internal.h>
|
|
#include <lwp_arch.h>
|
|
#include <lwp_user_mm.h>
|
|
#include <page.h>
|
|
|
|
#include <cpuport.h>
|
|
#include <encoding.h>
|
|
#include <stack.h>
|
|
#include <cache.h>
|
|
|
|
extern rt_ubase_t MMUTable[];
|
|
|
|
void *lwp_copy_return_code_to_user_stack()
|
|
{
|
|
void lwp_thread_return();
|
|
void lwp_thread_return_end();
|
|
rt_thread_t tid = rt_thread_self();
|
|
|
|
if (tid->user_stack != RT_NULL)
|
|
{
|
|
rt_size_t size = (rt_size_t)lwp_thread_return_end - (rt_size_t)lwp_thread_return;
|
|
rt_size_t userstack = (rt_size_t)tid->user_stack + tid->user_stack_size - size;
|
|
lwp_memcpy((void *)userstack, lwp_thread_return, size);
|
|
return (void *)userstack;
|
|
}
|
|
|
|
return RT_NULL;
|
|
}
|
|
|
|
rt_ubase_t lwp_fix_sp(rt_ubase_t cursp)
|
|
{
|
|
void lwp_thread_return();
|
|
void lwp_thread_return_end();
|
|
|
|
if (cursp == 0)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
return cursp - ((rt_size_t)lwp_thread_return_end - (rt_size_t)lwp_thread_return);
|
|
}
|
|
|
|
rt_thread_t rt_thread_sp_to_thread(void *spmember_addr)
|
|
{
|
|
return (rt_thread_t)(((rt_ubase_t)spmember_addr) - (offsetof(struct rt_thread, sp)));
|
|
}
|
|
|
|
void *get_thread_kernel_stack_top(rt_thread_t thread)
|
|
{
|
|
return (void *)(((rt_size_t)thread->stack_addr) + ((rt_size_t)thread->stack_size));
|
|
}
|
|
|
|
void *arch_get_user_sp(void)
|
|
{
|
|
/* user sp saved in interrupt context */
|
|
rt_thread_t self = rt_thread_self();
|
|
rt_uint8_t *stack_top = (rt_uint8_t *)self->stack_addr + self->stack_size;
|
|
struct rt_hw_stack_frame *frame = (struct rt_hw_stack_frame *)(stack_top - sizeof(struct rt_hw_stack_frame));
|
|
|
|
return (void *)frame->user_sp_exc_stack;
|
|
}
|
|
|
|
int arch_user_space_init(struct rt_lwp *lwp)
|
|
{
|
|
rt_ubase_t *mmu_table;
|
|
|
|
mmu_table = rt_hw_mmu_pgtbl_create();
|
|
if (!mmu_table)
|
|
{
|
|
return -RT_ENOMEM;
|
|
}
|
|
|
|
lwp->end_heap = USER_HEAP_VADDR;
|
|
lwp->aspace = rt_aspace_create(
|
|
(void *)USER_VADDR_START, USER_VADDR_TOP - USER_VADDR_START, mmu_table);
|
|
if (!lwp->aspace)
|
|
{
|
|
return -RT_ERROR;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void *arch_kernel_mmu_table_get(void)
|
|
{
|
|
return (void *)((char *)MMUTable);
|
|
}
|
|
|
|
void arch_user_space_free(struct rt_lwp *lwp)
|
|
{
|
|
if (lwp)
|
|
{
|
|
RT_ASSERT(lwp->aspace);
|
|
|
|
void *pgtbl = lwp->aspace->page_table;
|
|
rt_aspace_delete(lwp->aspace);
|
|
|
|
/* must be freed after aspace delete, pgtbl is required for unmap */
|
|
rt_hw_mmu_pgtbl_delete(pgtbl);
|
|
lwp->aspace = RT_NULL;
|
|
}
|
|
else
|
|
{
|
|
LOG_W("%s: NULL lwp as parameter", __func__);
|
|
RT_ASSERT(0);
|
|
}
|
|
}
|
|
|
|
long _sys_clone(void *arg[]);
|
|
long sys_clone(void *arg[])
|
|
{
|
|
return _sys_clone(arg);
|
|
}
|
|
|
|
long _sys_fork(void);
|
|
long sys_fork(void)
|
|
{
|
|
return _sys_fork();
|
|
}
|
|
|
|
long _sys_vfork(void);
|
|
long sys_vfork(void)
|
|
{
|
|
return _sys_fork();
|
|
}
|
|
|
|
/**
|
|
* set exec context for fork/clone.
|
|
*/
|
|
int arch_set_thread_context(void (*exit)(void), void *new_thread_stack,
|
|
void *user_stack, void **thread_sp)
|
|
{
|
|
RT_ASSERT(exit != RT_NULL);
|
|
RT_ASSERT(user_stack != RT_NULL);
|
|
RT_ASSERT(new_thread_stack != RT_NULL);
|
|
RT_ASSERT(thread_sp != RT_NULL);
|
|
struct rt_hw_stack_frame *syscall_frame;
|
|
struct rt_hw_stack_frame *thread_frame;
|
|
|
|
rt_uint8_t *stk;
|
|
rt_uint8_t *syscall_stk;
|
|
|
|
stk = (rt_uint8_t *)new_thread_stack;
|
|
/* reserve syscall context, all the registers are copyed from parent */
|
|
stk -= CTX_REG_NR * REGBYTES;
|
|
syscall_stk = stk;
|
|
|
|
syscall_frame = (struct rt_hw_stack_frame *)stk;
|
|
|
|
/* modify user sp */
|
|
syscall_frame->user_sp_exc_stack = (rt_ubase_t)user_stack;
|
|
|
|
/* skip ecall */
|
|
syscall_frame->epc += 4;
|
|
|
|
/* child return value is 0 */
|
|
syscall_frame->a0 = 0;
|
|
syscall_frame->a1 = 0;
|
|
|
|
/* reset thread area */
|
|
rt_thread_t thread = rt_container_of((unsigned long)thread_sp, struct rt_thread, sp);
|
|
syscall_frame->tp = (rt_ubase_t)thread->thread_idr;
|
|
|
|
#ifdef ARCH_USING_NEW_CTX_SWITCH
|
|
extern void *_rt_hw_stack_init(rt_ubase_t *sp, rt_ubase_t ra, rt_ubase_t sstatus);
|
|
rt_ubase_t sstatus = read_csr(sstatus) | SSTATUS_SPP;
|
|
sstatus &= ~SSTATUS_SIE;
|
|
|
|
/* compatible to RESTORE_CONTEXT */
|
|
stk = (void *)_rt_hw_stack_init((rt_ubase_t *)stk, (rt_ubase_t)exit, sstatus);
|
|
#else
|
|
/* build temp thread context */
|
|
stk -= sizeof(struct rt_hw_stack_frame);
|
|
|
|
thread_frame = (struct rt_hw_stack_frame *)stk;
|
|
|
|
int i;
|
|
for (i = 0; i < sizeof(struct rt_hw_stack_frame) / sizeof(rt_ubase_t); i++)
|
|
{
|
|
((rt_ubase_t *)thread_frame)[i] = 0xdeadbeaf;
|
|
}
|
|
|
|
/* set pc for thread */
|
|
thread_frame->epc = (rt_ubase_t)exit;
|
|
|
|
/* set old exception mode as supervisor, because in kernel */
|
|
thread_frame->sstatus = read_csr(sstatus) | SSTATUS_SPP;
|
|
thread_frame->sstatus &= ~SSTATUS_SIE; /* must disable interrupt */
|
|
|
|
/* set stack as syscall stack */
|
|
thread_frame->user_sp_exc_stack = (rt_ubase_t)syscall_stk;
|
|
|
|
#endif /* ARCH_USING_NEW_CTX_SWITCH */
|
|
/* save new stack top */
|
|
*thread_sp = (void *)stk;
|
|
|
|
/**
|
|
* The stack for child thread:
|
|
*
|
|
* +------------------------+ --> kernel stack top
|
|
* | syscall stack |
|
|
* | |
|
|
* | @sp | --> `user_stack`
|
|
* | @epc | --> user ecall addr + 4 (skip ecall)
|
|
* | @a0&a1 | --> 0 (for child return 0)
|
|
* | |
|
|
* +------------------------+ --> temp thread stack top
|
|
* | temp thread stack | ^
|
|
* | | |
|
|
* | @sp | ---------/
|
|
* | @epc | --> `exit` (arch_clone_exit/arch_fork_exit)
|
|
* | |
|
|
* +------------------------+ --> thread sp
|
|
*/
|
|
}
|
|
|
|
#define ALGIN_BYTES (16)
|
|
|
|
struct signal_ucontext
|
|
{
|
|
rt_int64_t sigreturn;
|
|
lwp_sigset_t save_sigmask;
|
|
|
|
siginfo_t si;
|
|
|
|
rt_align(16)
|
|
struct rt_hw_stack_frame frame;
|
|
};
|
|
|
|
void *arch_signal_ucontext_restore(rt_base_t user_sp)
|
|
{
|
|
struct signal_ucontext *new_sp;
|
|
new_sp = (void *)user_sp;
|
|
|
|
if (lwp_user_accessable(new_sp, sizeof(*new_sp)))
|
|
{
|
|
lwp_thread_signal_mask(rt_thread_self(), LWP_SIG_MASK_CMD_SET_MASK, &new_sp->save_sigmask, RT_NULL);
|
|
}
|
|
else
|
|
{
|
|
LOG_I("User frame corrupted during signal handling\nexiting...");
|
|
sys_exit_group(EXIT_FAILURE);
|
|
}
|
|
|
|
return (void *)&new_sp->frame;
|
|
}
|
|
|
|
void *arch_signal_ucontext_save(int signo, siginfo_t *psiginfo,
|
|
struct rt_hw_stack_frame *exp_frame, rt_base_t user_sp,
|
|
lwp_sigset_t *save_sig_mask)
|
|
{
|
|
struct signal_ucontext *new_sp;
|
|
new_sp = (void *)(user_sp - sizeof(struct signal_ucontext));
|
|
|
|
if (lwp_user_accessable(new_sp, sizeof(*new_sp)))
|
|
{
|
|
/* push psiginfo */
|
|
if (psiginfo)
|
|
{
|
|
lwp_memcpy(&new_sp->si, psiginfo, sizeof(*psiginfo));
|
|
}
|
|
|
|
lwp_memcpy(&new_sp->frame, exp_frame, sizeof(*exp_frame));
|
|
|
|
/* copy the save_sig_mask */
|
|
lwp_memcpy(&new_sp->save_sigmask, save_sig_mask, sizeof(lwp_sigset_t));
|
|
|
|
/* copy lwp_sigreturn */
|
|
const size_t lwp_sigreturn_bytes = 8;
|
|
extern void lwp_sigreturn(void);
|
|
/* -> ensure that the sigreturn start at the outer most boundary */
|
|
lwp_memcpy(&new_sp->sigreturn, &lwp_sigreturn, lwp_sigreturn_bytes);
|
|
|
|
/**
|
|
* synchronize dcache & icache if target is
|
|
* a Harvard Architecture machine, otherwise
|
|
* do nothing
|
|
*/
|
|
rt_hw_sync_cache_local(&new_sp->sigreturn, 8);
|
|
}
|
|
else
|
|
{
|
|
LOG_I("%s: User stack overflow", __func__);
|
|
sys_exit_group(EXIT_FAILURE);
|
|
}
|
|
|
|
return new_sp;
|
|
}
|
|
|
|
/**
|
|
* void lwp_exec_user(void *args, void *kernel_stack, void *user_entry)
|
|
*/
|
|
void lwp_exec_user(void *args, void *kernel_stack, void *user_entry)
|
|
{
|
|
arch_start_umode(args, user_entry, (void *)USER_STACK_VEND, kernel_stack);
|
|
}
|
|
|
|
#endif /* ARCH_MM_MMU */
|
|
|
|
int arch_backtrace_uthread(rt_thread_t thread)
|
|
{
|
|
struct rt_hw_backtrace_frame frame;
|
|
struct rt_hw_stack_frame *stack;
|
|
|
|
if (thread && thread->lwp)
|
|
{
|
|
stack = thread->user_ctx.ctx;
|
|
if ((long)stack > (unsigned long)thread->stack_addr
|
|
&& (long)stack < (unsigned long)thread->stack_addr + thread->stack_size)
|
|
{
|
|
frame.pc = stack->epc;
|
|
frame.fp = stack->s0_fp;
|
|
lwp_backtrace_frame(thread, &frame);
|
|
return 0;
|
|
}
|
|
else
|
|
return -1;
|
|
}
|
|
return -1;
|
|
}
|