/* * Copyright (c) 2006-2021, RT-Thread Development Team * * SPDX-License-Identifier: Apache-2.0 * * Change Logs: * Date Author Notes * 2020-11-18 Jesven first version * 2021-02-03 lizhirui port to riscv64 * 2021-02-06 lizhirui add thread filter * 2021-02-19 lizhirui port to new version of rt-smart * 2021-03-02 lizhirui add a auxillary function for interrupt * 2021-03-04 lizhirui delete thread filter * 2021-03-04 lizhirui modify for new version of rt-smart * 2021-11-22 JasonHu add lwp_set_thread_context * 2021-11-30 JasonHu add clone/fork support * 2023-07-16 Shell Move part of the codes to C from asm in signal handling */ #include #include #include #ifdef ARCH_MM_MMU #define DBG_TAG "lwp.arch" #define DBG_LVL DBG_INFO #include #include #include #include #include #include #include #include #include extern rt_ubase_t MMUTable[]; void *lwp_copy_return_code_to_user_stack() { void lwp_thread_return(); void lwp_thread_return_end(); rt_thread_t tid = rt_thread_self(); if (tid->user_stack != RT_NULL) { rt_size_t size = (rt_size_t)lwp_thread_return_end - (rt_size_t)lwp_thread_return; rt_size_t userstack = (rt_size_t)tid->user_stack + tid->user_stack_size - size; lwp_memcpy((void *)userstack, lwp_thread_return, size); return (void *)userstack; } return RT_NULL; } rt_ubase_t lwp_fix_sp(rt_ubase_t cursp) { void lwp_thread_return(); void lwp_thread_return_end(); if (cursp == 0) { return 0; } return cursp - ((rt_size_t)lwp_thread_return_end - (rt_size_t)lwp_thread_return); } rt_thread_t rt_thread_sp_to_thread(void *spmember_addr) { return (rt_thread_t)(((rt_ubase_t)spmember_addr) - (offsetof(struct rt_thread, sp))); } void *get_thread_kernel_stack_top(rt_thread_t thread) { return (void *)(((rt_size_t)thread->stack_addr) + ((rt_size_t)thread->stack_size)); } void *arch_get_user_sp(void) { /* user sp saved in interrupt context */ rt_thread_t self = rt_thread_self(); rt_uint8_t *stack_top = (rt_uint8_t *)self->stack_addr + self->stack_size; struct rt_hw_stack_frame *frame = (struct rt_hw_stack_frame *)(stack_top - sizeof(struct rt_hw_stack_frame)); return (void *)frame->user_sp_exc_stack; } int arch_user_space_init(struct rt_lwp *lwp) { rt_ubase_t *mmu_table; mmu_table = (rt_ubase_t *)rt_pages_alloc_ext(0, PAGE_ANY_AVAILABLE); if (!mmu_table) { return -RT_ENOMEM; } lwp->end_heap = USER_HEAP_VADDR; rt_memcpy(mmu_table, rt_kernel_space.page_table, ARCH_PAGE_SIZE); rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, mmu_table, ARCH_PAGE_SIZE); lwp->aspace = rt_aspace_create( (void *)USER_VADDR_START, USER_VADDR_TOP - USER_VADDR_START, mmu_table); if (!lwp->aspace) { return -RT_ERROR; } return 0; } void *arch_kernel_mmu_table_get(void) { return (void *)((char *)MMUTable); } void arch_user_space_free(struct rt_lwp *lwp) { if (lwp) { RT_ASSERT(lwp->aspace); void *pgtbl = lwp->aspace->page_table; rt_aspace_delete(lwp->aspace); /* must be freed after aspace delete, pgtbl is required for unmap */ rt_pages_free(pgtbl, 0); lwp->aspace = RT_NULL; } else { LOG_W("%s: NULL lwp as parameter", __func__); RT_ASSERT(0); } } long _sys_clone(void *arg[]); long sys_clone(void *arg[]) { return _sys_clone(arg); } long _sys_fork(void); long sys_fork(void) { return _sys_fork(); } long _sys_vfork(void); long sys_vfork(void) { return _sys_fork(); } /** * set exec context for fork/clone. */ int arch_set_thread_context(void (*exit)(void), void *new_thread_stack, void *user_stack, void **thread_sp) { RT_ASSERT(exit != RT_NULL); RT_ASSERT(user_stack != RT_NULL); RT_ASSERT(new_thread_stack != RT_NULL); RT_ASSERT(thread_sp != RT_NULL); struct rt_hw_stack_frame *syscall_frame; struct rt_hw_stack_frame *thread_frame; rt_uint8_t *stk; rt_uint8_t *syscall_stk; stk = (rt_uint8_t *)new_thread_stack; /* reserve syscall context, all the registers are copyed from parent */ stk -= CTX_REG_NR * REGBYTES; syscall_stk = stk; syscall_frame = (struct rt_hw_stack_frame *)stk; /* modify user sp */ syscall_frame->user_sp_exc_stack = (rt_ubase_t)user_stack; /* skip ecall */ syscall_frame->epc += 4; /* child return value is 0 */ syscall_frame->a0 = 0; syscall_frame->a1 = 0; /* reset thread area */ rt_thread_t thread = rt_container_of((unsigned long)thread_sp, struct rt_thread, sp); syscall_frame->tp = (rt_ubase_t)thread->thread_idr; #ifdef ARCH_USING_NEW_CTX_SWITCH extern void *_rt_hw_stack_init(rt_ubase_t *sp, rt_ubase_t ra, rt_ubase_t sstatus); rt_ubase_t sstatus = read_csr(sstatus) | SSTATUS_SPP; sstatus &= ~SSTATUS_SIE; /* compatible to RESTORE_CONTEXT */ stk = (void *)_rt_hw_stack_init((rt_ubase_t *)stk, (rt_ubase_t)exit, sstatus); #else /* build temp thread context */ stk -= sizeof(struct rt_hw_stack_frame); thread_frame = (struct rt_hw_stack_frame *)stk; int i; for (i = 0; i < sizeof(struct rt_hw_stack_frame) / sizeof(rt_ubase_t); i++) { ((rt_ubase_t *)thread_frame)[i] = 0xdeadbeaf; } /* set pc for thread */ thread_frame->epc = (rt_ubase_t)exit; /* set old exception mode as supervisor, because in kernel */ thread_frame->sstatus = read_csr(sstatus) | SSTATUS_SPP; thread_frame->sstatus &= ~SSTATUS_SIE; /* must disable interrupt */ /* set stack as syscall stack */ thread_frame->user_sp_exc_stack = (rt_ubase_t)syscall_stk; #endif /* ARCH_USING_NEW_CTX_SWITCH */ /* save new stack top */ *thread_sp = (void *)stk; /** * The stack for child thread: * * +------------------------+ --> kernel stack top * | syscall stack | * | | * | @sp | --> `user_stack` * | @epc | --> user ecall addr + 4 (skip ecall) * | @a0&a1 | --> 0 (for child return 0) * | | * +------------------------+ --> temp thread stack top * | temp thread stack | ^ * | | | * | @sp | ---------/ * | @epc | --> `exit` (arch_clone_exit/arch_fork_exit) * | | * +------------------------+ --> thread sp */ } #define ALGIN_BYTES (16) struct signal_ucontext { rt_int64_t sigreturn; lwp_sigset_t save_sigmask; siginfo_t si; rt_align(16) struct rt_hw_stack_frame frame; }; void *arch_signal_ucontext_restore(rt_base_t user_sp) { struct signal_ucontext *new_sp; new_sp = (void *)user_sp; if (lwp_user_accessable(new_sp, sizeof(*new_sp))) { lwp_thread_signal_mask(rt_thread_self(), LWP_SIG_MASK_CMD_SET_MASK, &new_sp->save_sigmask, RT_NULL); } else { LOG_I("User frame corrupted during signal handling\nexiting..."); sys_exit_group(EXIT_FAILURE); } return (void *)&new_sp->frame; } void *arch_signal_ucontext_save(int signo, siginfo_t *psiginfo, struct rt_hw_stack_frame *exp_frame, rt_base_t user_sp, lwp_sigset_t *save_sig_mask) { struct signal_ucontext *new_sp; new_sp = (void *)(user_sp - sizeof(struct signal_ucontext)); if (lwp_user_accessable(new_sp, sizeof(*new_sp))) { /* push psiginfo */ if (psiginfo) { lwp_memcpy(&new_sp->si, psiginfo, sizeof(*psiginfo)); } lwp_memcpy(&new_sp->frame, exp_frame, sizeof(*exp_frame)); /* copy the save_sig_mask */ lwp_memcpy(&new_sp->save_sigmask, save_sig_mask, sizeof(lwp_sigset_t)); /* copy lwp_sigreturn */ const size_t lwp_sigreturn_bytes = 8; extern void lwp_sigreturn(void); /* -> ensure that the sigreturn start at the outer most boundary */ lwp_memcpy(&new_sp->sigreturn, &lwp_sigreturn, lwp_sigreturn_bytes); /** * synchronize dcache & icache if target is * a Harvard Architecture machine, otherwise * do nothing */ rt_hw_sync_cache_local(&new_sp->sigreturn, 8); } else { LOG_I("%s: User stack overflow", __func__); sys_exit_group(EXIT_FAILURE); } return new_sp; } /** * void lwp_exec_user(void *args, void *kernel_stack, void *user_entry) */ void lwp_exec_user(void *args, void *kernel_stack, void *user_entry) { arch_start_umode(args, user_entry, (void *)USER_STACK_VEND, kernel_stack); } #endif /* ARCH_MM_MMU */