/* * Copyright (c) 2006-2021, RT-Thread Development Team * * SPDX-License-Identifier: Apache-2.0 * * Change Logs: * Date Author Notes * 2006-03-12 Bernard first version * 2018-11-02 heyuanjie fix complie error in iar * 2021-02-03 lizhirui add 64-bit arch support and riscv64 arch support * 2021-08-26 linzhenxing add lwp_setcwd\lwp_getcwd * 2023-02-20 wangxiaoyao inv icache before new app startup * 2023-02-20 wangxiaoyao fix bug on foreground app switch */ #define DBG_TAG "LWP" #define DBG_LVL DBG_WARNING #include #include #include #include #include #include /* rename() */ #include #include #include /* statfs() */ #include #ifndef RT_USING_DFS #error "lwp need file system(RT_USING_DFS)" #endif #include "lwp.h" #include "lwp_arch.h" #include "lwp_arch_comm.h" #include "lwp_signal.h" #include "lwp_dbg.h" #include "console.h" #ifdef ARCH_MM_MMU #include #endif /* end of ARCH_MM_MMU */ #ifndef O_DIRECTORY #define O_DIRECTORY 0x200000 #endif #ifndef O_BINARY #define O_BINARY 0x10000 #endif static const char elf_magic[] = {0x7f, 'E', 'L', 'F'}; #ifdef DFS_USING_WORKDIR extern char working_directory[]; #endif static struct termios stdin_termios, old_stdin_termios; int load_ldso(struct rt_lwp *lwp, char *exec_name, char *const argv[], char *const envp[]); struct termios *get_old_termios(void) { return &old_stdin_termios; } void lwp_setcwd(char *buf) { struct rt_lwp *lwp = RT_NULL; if(strlen(buf) >= DFS_PATH_MAX) { rt_kprintf("buf too long!\n"); return ; } lwp = (struct rt_lwp *)rt_thread_self()->lwp; if (lwp) { rt_strncpy(lwp->working_directory, buf, DFS_PATH_MAX); } else { rt_strncpy(working_directory, buf, DFS_PATH_MAX); } return ; } char *lwp_getcwd(void) { char *dir_buf = RT_NULL; struct rt_lwp *lwp = RT_NULL; lwp = (struct rt_lwp *)rt_thread_self()->lwp; if (lwp) { if(lwp->working_directory[0] != '/') { dir_buf = &working_directory[0]; } else { dir_buf = &lwp->working_directory[0]; } } else dir_buf = &working_directory[0]; return dir_buf; } /** * RT-Thread light-weight process */ void lwp_set_kernel_sp(uint32_t *sp) { rt_thread_self()->kernel_sp = (rt_uint32_t *)sp; } uint32_t *lwp_get_kernel_sp(void) { #ifdef ARCH_MM_MMU return (uint32_t *)rt_thread_self()->sp; #else uint32_t* kernel_sp; extern rt_uint32_t rt_interrupt_from_thread; extern rt_uint32_t rt_thread_switch_interrupt_flag; if (rt_thread_switch_interrupt_flag) { kernel_sp = (uint32_t *)((rt_thread_t)rt_container_of(rt_interrupt_from_thread, struct rt_thread, sp))->kernel_sp; } else { kernel_sp = (uint32_t *)rt_thread_self()->kernel_sp; } return kernel_sp; #endif } #ifdef ARCH_MM_MMU struct process_aux *lwp_argscopy(struct rt_lwp *lwp, int argc, char **argv, char **envp) { int size = sizeof(size_t) * 5; /* store argc, argv, envp, aux, NULL */ int *args; char *str; char *str_k; char **new_argve; int i; int len; size_t *args_k; struct process_aux *aux; for (i = 0; i < argc; i++) { size += (rt_strlen(argv[i]) + 1); } size += (sizeof(size_t) * argc); i = 0; if (envp) { while (envp[i] != 0) { size += (rt_strlen(envp[i]) + 1); size += sizeof(size_t); i++; } } /* for aux */ size += sizeof(struct process_aux); if (size > ARCH_PAGE_SIZE) { return RT_NULL; } /* args = (int *)lwp_map_user(lwp, 0, size); */ args = (int *)lwp_map_user(lwp, (void *)(USER_VADDR_TOP - ARCH_PAGE_SIZE), size, 0); if (args == RT_NULL) { return RT_NULL; } args_k = (size_t *)lwp_v2p(lwp, args); args_k = (size_t *)((size_t)args_k - PV_OFFSET); /* argc, argv[], 0, envp[], 0 , aux[] */ str = (char *)((size_t)args + (argc + 2 + i + 1 + AUX_ARRAY_ITEMS_NR * 2 + 1) * sizeof(size_t)); str_k = (char *)((size_t)args_k + (argc + 2 + i + 1 + AUX_ARRAY_ITEMS_NR * 2 + 1) * sizeof(size_t)); new_argve = (char **)&args_k[1]; args_k[0] = argc; for (i = 0; i < argc; i++) { len = rt_strlen(argv[i]) + 1; new_argve[i] = str; rt_memcpy(str_k, argv[i], len); str += len; str_k += len; } new_argve[i] = 0; i++; new_argve[i] = 0; if (envp) { int j; for (j = 0; envp[j] != 0; j++) { len = rt_strlen(envp[j]) + 1; new_argve[i] = str; rt_memcpy(str_k, envp[j], len); str += len; str_k += len; i++; } new_argve[i] = 0; } i++; /* aux */ aux = (struct process_aux *)(new_argve + i); aux->item[0].key = AT_EXECFN; aux->item[0].value = (size_t)(size_t)new_argve[0]; i += AUX_ARRAY_ITEMS_NR * 2; new_argve[i] = 0; rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, args_k, size); lwp->args = args; return aux; } #else static struct process_aux *lwp_argscopy(struct rt_lwp *lwp, int argc, char **argv, char **envp) { #ifdef ARCH_MM_MMU int size = sizeof(int) * 5; /* store argc, argv, envp, aux, NULL */ struct process_aux *aux; #else int size = sizeof(int) * 4; /* store argc, argv, envp, NULL */ #endif /* ARCH_MM_MMU */ int *args; char *str; char **new_argve; int i; int len; for (i = 0; i < argc; i++) { size += (rt_strlen(argv[i]) + 1); } size += (sizeof(int) * argc); i = 0; if (envp) { while (envp[i] != 0) { size += (rt_strlen(envp[i]) + 1); size += sizeof(int); i++; } } #ifdef ARCH_MM_MMU /* for aux */ size += sizeof(struct process_aux); args = (int *)rt_malloc(size); if (args == RT_NULL) { return RT_NULL; } /* argc, argv[], 0, envp[], 0 */ str = (char *)((size_t)args + (argc + 2 + i + 1 + AUX_ARRAY_ITEMS_NR * 2 + 1) * sizeof(int)); #else args = (int *)rt_malloc(size); if (args == RT_NULL) { return RT_NULL; } str = (char*)((int)args + (argc + 2 + i + 1) * sizeof(int)); #endif /* ARCH_MM_MMU */ new_argve = (char **)&args[1]; args[0] = argc; for (i = 0; i < argc; i++) { len = rt_strlen(argv[i]) + 1; new_argve[i] = str; rt_memcpy(str, argv[i], len); str += len; } new_argve[i] = 0; i++; new_argve[i] = 0; if (envp) { int j; for (j = 0; envp[j] != 0; j++) { len = rt_strlen(envp[j]) + 1; new_argve[i] = str; rt_memcpy(str, envp[j], len); str += len; i++; } new_argve[i] = 0; } #ifdef ARCH_MM_MMU /* aux */ aux = (struct process_aux *)(new_argve + i); aux->item[0].key = AT_EXECFN; aux->item[0].value = (uint32_t)(size_t)new_argve[0]; i += AUX_ARRAY_ITEMS_NR * 2; new_argve[i] = 0; lwp->args = args; return aux; #else lwp->args = args; lwp->args_length = size; return (struct process_aux *)(new_argve + i); #endif /* ARCH_MM_MMU */ } #endif #ifdef ARCH_MM_MMU #define check_off(voff, vlen) \ do \ { \ if (voff > vlen) \ { \ result = -RT_ERROR; \ goto _exit; \ } \ } while (0) #define check_read(vrlen, vrlen_want) \ do \ { \ if (vrlen < vrlen_want) \ { \ result = -RT_ERROR; \ goto _exit; \ } \ } while (0) static size_t load_fread(void *ptr, size_t size, size_t nmemb, int fd) { size_t read_block = 0; while (nmemb) { size_t count; count = read(fd, ptr, size * nmemb) / size; if (count < nmemb) { LOG_E("ERROR: file size error!"); break; } ptr = (void *)((uint8_t *)ptr + (count * size)); nmemb -= count; read_block += count; } return read_block; } typedef struct { Elf_Word st_name; Elf_Addr st_value; Elf_Word st_size; unsigned char st_info; unsigned char st_other; Elf_Half st_shndx; } Elf_sym; #ifdef ARCH_MM_MMU struct map_range { void *start; size_t size; }; static void expand_map_range(struct map_range *m, void *start, size_t size) { if (!m->start) { m->start = start; m->size = size; } else { void *end = (void *)((char*)start + size); void *mend = (void *)((char*)m->start + m->size); if (m->start > start) { m->start = start; } if (mend < end) { mend = end; } m->size = (char *)mend - (char *)m->start; } } static int map_range_ckeck(struct map_range *m1, struct map_range *m2) { void *m1_start = (void *)((size_t)m1->start & ~ARCH_PAGE_MASK); void *m1_end = (void *)((((size_t)m1->start + m1->size) + ARCH_PAGE_MASK) & ~ARCH_PAGE_MASK); void *m2_start = (void *)((size_t)m2->start & ~ARCH_PAGE_MASK); void *m2_end = (void *)((((size_t)m2->start + m2->size) + ARCH_PAGE_MASK) & ~ARCH_PAGE_MASK); if (m1->size) { if (m1_start < (void *)USER_LOAD_VADDR) { return -1; } if (m1_start > (void *)USER_STACK_VSTART) { return -1; } if (m1_end < (void *)USER_LOAD_VADDR) { return -1; } if (m1_end > (void *)USER_STACK_VSTART) { return -1; } } if (m2->size) { if (m2_start < (void *)USER_LOAD_VADDR) { return -1; } if (m2_start > (void *)USER_STACK_VSTART) { return -1; } if (m2_end < (void *)USER_LOAD_VADDR) { return -1; } if (m2_end > (void *)USER_STACK_VSTART) { return -1; } } if ((m1->size != 0) && (m2->size != 0)) { if (m1_start < m2_start) { if (m1_end > m2_start) { return -1; } } else /* m2_start <= m1_start */ { if (m2_end > m1_start) { return -1; } } } return 0; } #endif static int load_elf(int fd, int len, struct rt_lwp *lwp, uint8_t *load_addr, struct process_aux *aux) { uint32_t i; uint32_t off = 0; size_t load_off = 0; char *p_section_str = 0; Elf_sym *dynsym = 0; Elf_Ehdr eheader; Elf_Phdr pheader; Elf_Shdr sheader; int result = RT_EOK; uint32_t magic; size_t read_len; void *got_start = 0; size_t got_size = 0; void *rel_dyn_start = 0; size_t rel_dyn_size = 0; size_t dynsym_off = 0; size_t dynsym_size = 0; #ifdef ARCH_MM_MMU struct map_range user_area[2] = {{NULL, 0}, {NULL, 0}}; /* 0 is text, 1 is data */ void *pa, *va; void *va_self; #endif if (len < sizeof eheader) { LOG_E("len < sizeof eheader!"); return -RT_ERROR; } lseek(fd, 0, SEEK_SET); read_len = load_fread(&magic, 1, sizeof magic, fd); check_read(read_len, sizeof magic); if (memcmp(elf_magic, &magic, 4) != 0) { LOG_E("elf_magic not same, magic:0x%x!", magic); return -RT_ERROR; } lseek(fd, off, SEEK_SET); read_len = load_fread(&eheader, 1, sizeof eheader, fd); check_read(read_len, sizeof eheader); #ifndef ARCH_CPU_64BIT if (eheader.e_ident[4] != 1) { /* not 32bit */ LOG_E("elf not 32bit, %d!", eheader.e_ident[4]); return -RT_ERROR; } #else if (eheader.e_ident[4] != 2) { /* not 64bit */ LOG_E("elf not 64bit, %d!", eheader.e_ident[4]); return -RT_ERROR; } #endif if (eheader.e_ident[6] != 1) { /* ver not 1 */ LOG_E("elf Version not 1,ver:%d!", eheader.e_ident[6]); return -RT_ERROR; } if ((eheader.e_type != ET_DYN) #ifdef ARCH_MM_MMU && (eheader.e_type != ET_EXEC) #endif ) { /* not pie or exec elf */ LOG_E("elf type not pie or exec, type:%d!", eheader.e_type); return -RT_ERROR; } #ifdef ARCH_MM_MMU { off = eheader.e_phoff; for (i = 0; i < eheader.e_phnum; i++, off += sizeof pheader) { check_off(off, len); lseek(fd, off, SEEK_SET); read_len = load_fread(&pheader, 1, sizeof pheader, fd); check_read(read_len, sizeof pheader); if (pheader.p_type == PT_DYNAMIC) { /* load ld.so */ return 1; /* 1 means dynamic */ } } } #endif if (eheader.e_entry != 0) { if ((eheader.e_entry != USER_LOAD_VADDR) && (eheader.e_entry != LDSO_LOAD_VADDR)) { /* the entry is invalidate */ LOG_E("elf entry is invalidate, entry:0x%x!", eheader.e_entry); return -RT_ERROR; } } { /* load aux */ uint8_t *process_header; size_t process_header_size; off = eheader.e_phoff; process_header_size = eheader.e_phnum * sizeof pheader; #ifdef ARCH_MM_MMU if (process_header_size > ARCH_PAGE_SIZE - sizeof(char[16])) { LOG_E("process_header_size too big, size:0x%x!", process_header_size); return -RT_ERROR; } va = (uint8_t *)lwp_map_user(lwp, (void *)(USER_VADDR_TOP - ARCH_PAGE_SIZE * 2), process_header_size, 0); if (!va) { LOG_E("lwp map user failed!"); return -RT_ERROR; } pa = lwp_v2p(lwp, va); process_header = (uint8_t *)pa - PV_OFFSET; #else process_header = (uint8_t *)rt_malloc(process_header_size + sizeof(char[16])); if (!process_header) { LOG_E("process_header malloc failed, size:0x%x!", process_header_size + sizeof(char[16])); return -RT_ERROR; } #endif check_off(off, len); lseek(fd, off, SEEK_SET); read_len = load_fread(process_header, 1, process_header_size, fd); check_read(read_len, process_header_size); #ifdef ARCH_MM_MMU rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, process_header, process_header_size); #endif aux->item[1].key = AT_PAGESZ; #ifdef ARCH_MM_MMU aux->item[1].value = ARCH_PAGE_SIZE; #else aux->item[1].value = RT_MM_PAGE_SIZE; #endif aux->item[2].key = AT_RANDOM; { uint32_t random_value = rt_tick_get(); uint8_t *random; #ifdef ARCH_MM_MMU uint8_t *krandom; random = (uint8_t *)(USER_VADDR_TOP - ARCH_PAGE_SIZE - sizeof(char[16])); krandom = (uint8_t *)lwp_v2p(lwp, random); krandom = (uint8_t *)krandom - PV_OFFSET; rt_memcpy(krandom, &random_value, sizeof random_value); #else random = (uint8_t *)(process_header + process_header_size); rt_memcpy(random, &random_value, sizeof random_value); #endif aux->item[2].value = (size_t)random; } aux->item[3].key = AT_PHDR; #ifdef ARCH_MM_MMU aux->item[3].value = (size_t)va; #else aux->item[3].value = (size_t)process_header; #endif aux->item[4].key = AT_PHNUM; aux->item[4].value = eheader.e_phnum; aux->item[5].key = AT_PHENT; aux->item[5].value = sizeof pheader; #ifdef ARCH_MM_MMU rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, aux, sizeof *aux); #endif } if (load_addr) { load_off = (size_t)load_addr; } #ifdef ARCH_MM_MMU else { /* map user */ off = eheader.e_shoff; for (i = 0; i < eheader.e_shnum; i++, off += sizeof sheader) { check_off(off, len); lseek(fd, off, SEEK_SET); read_len = load_fread(&sheader, 1, sizeof sheader, fd); check_read(read_len, sizeof sheader); if ((sheader.sh_flags & SHF_ALLOC) == 0) { continue; } switch (sheader.sh_type) { case SHT_PROGBITS: if ((sheader.sh_flags & SHF_WRITE) == 0) { expand_map_range(&user_area[0], (void *)sheader.sh_addr, sheader.sh_size); } else { expand_map_range(&user_area[1], (void *)sheader.sh_addr, sheader.sh_size); } break; case SHT_NOBITS: expand_map_range(&user_area[1], (void *)sheader.sh_addr, sheader.sh_size); break; default: expand_map_range(&user_area[1], (void *)sheader.sh_addr, sheader.sh_size); break; } } if (user_area[0].size == 0) { /* no code */ result = -RT_ERROR; goto _exit; } if (user_area[0].start == NULL) { /* DYN */ load_off = USER_LOAD_VADDR; user_area[0].start = (void *)((char*)user_area[0].start + load_off); user_area[1].start = (void *)((char*)user_area[1].start + load_off); } if (map_range_ckeck(&user_area[0], &user_area[1]) != 0) { result = -RT_ERROR; goto _exit; } /* text and data */ for (i = 0; i < 2; i++) { if (user_area[i].size != 0) { va = lwp_map_user(lwp, user_area[i].start, user_area[i].size, (i == 0)); if (!va || (va != user_area[i].start)) { result = -RT_ERROR; goto _exit; } } } lwp->text_size = user_area[0].size; } #else else { size_t start = -1UL; size_t end = 0UL; size_t total_size; off = eheader.e_shoff; for (i = 0; i < eheader.e_shnum; i++, off += sizeof sheader) { check_off(off, len); lseek(fd, off, SEEK_SET); read_len = load_fread(&sheader, 1, sizeof sheader, fd); check_read(read_len, sizeof sheader); if ((sheader.sh_flags & SHF_ALLOC) == 0) { continue; } switch (sheader.sh_type) { case SHT_PROGBITS: case SHT_NOBITS: if (start > sheader.sh_addr) { start = sheader.sh_addr; } if (sheader.sh_addr + sheader.sh_size > end) { end = sheader.sh_addr + sheader.sh_size; } break; default: break; } } total_size = end - start; #ifdef RT_USING_CACHE load_off = (size_t)rt_malloc_align(total_size, RT_CPU_CACHE_LINE_SZ); #else load_off = (size_t)rt_malloc(total_size); #endif if (load_off == 0) { LOG_E("alloc text memory faild!"); result = -RT_ENOMEM; goto _exit; } else { LOG_D("lwp text malloc : %p, size: %d!", (void *)load_off, lwp->text_size); } lwp->load_off = load_off; /* for free */ lwp->text_size = total_size; } #endif lwp->text_entry = (void *)(eheader.e_entry + load_off); off = eheader.e_phoff; for (i = 0; i < eheader.e_phnum; i++, off += sizeof pheader) { check_off(off, len); lseek(fd, off, SEEK_SET); read_len = load_fread(&pheader, 1, sizeof pheader, fd); check_read(read_len, sizeof pheader); if (pheader.p_type == PT_LOAD) { if (pheader.p_filesz > pheader.p_memsz) { LOG_E("pheader.p_filesz > pheader.p_memsz, p_filesz:0x%x;p_memsz:0x%x!", pheader.p_filesz, pheader.p_memsz); return -RT_ERROR; } check_off(pheader.p_offset, len); lseek(fd, pheader.p_offset, SEEK_SET); #ifdef ARCH_MM_MMU { uint32_t size = pheader.p_filesz; size_t tmp_len = 0; va = (void *)(pheader.p_vaddr + load_addr); read_len = 0; while (size) { pa = lwp_v2p(lwp, va); va_self = (void *)((char *)pa - PV_OFFSET); LOG_D("va_self = %p pa = %p", va_self, pa); tmp_len = (size < ARCH_PAGE_SIZE) ? size : ARCH_PAGE_SIZE; tmp_len = load_fread(va_self, 1, tmp_len, fd); rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, va_self, tmp_len); read_len += tmp_len; size -= tmp_len; va = (void *)((char *)va + ARCH_PAGE_SIZE); } } #else read_len = load_fread((void*)(pheader.p_vaddr + load_off), 1, pheader.p_filesz, fd); #endif check_read(read_len, pheader.p_filesz); if (pheader.p_filesz < pheader.p_memsz) { #ifdef ARCH_MM_MMU uint32_t size = pheader.p_memsz - pheader.p_filesz; uint32_t size_s; uint32_t off; off = pheader.p_filesz & ARCH_PAGE_MASK; va = (void *)((pheader.p_vaddr + pheader.p_filesz + load_off) & ~ARCH_PAGE_MASK); while (size) { size_s = (size < ARCH_PAGE_SIZE - off) ? size : ARCH_PAGE_SIZE - off; pa = lwp_v2p(lwp, va); va_self = (void *)((char *)pa - PV_OFFSET); memset((void *)((char *)va_self + off), 0, size_s); rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, (void *)((char *)va_self + off), size_s); off = 0; size -= size_s; va = (void *)((char *)va + ARCH_PAGE_SIZE); } #else memset((uint8_t *)pheader.p_vaddr + pheader.p_filesz + load_off, 0, (size_t)(pheader.p_memsz - pheader.p_filesz)); #endif } } } /* relocate */ if (eheader.e_type == ET_DYN) { /* section info */ off = eheader.e_shoff; /* find section string table */ check_off(off, len); lseek(fd, off + (sizeof sheader) * eheader.e_shstrndx, SEEK_SET); read_len = load_fread(&sheader, 1, sizeof sheader, fd); check_read(read_len, sizeof sheader); p_section_str = (char *)rt_malloc(sheader.sh_size); if (!p_section_str) { LOG_E("out of memory!"); result = -ENOMEM; goto _exit; } check_off(sheader.sh_offset, len); lseek(fd, sheader.sh_offset, SEEK_SET); read_len = load_fread(p_section_str, 1, sheader.sh_size, fd); check_read(read_len, sheader.sh_size); check_off(off, len); lseek(fd, off, SEEK_SET); for (i = 0; i < eheader.e_shnum; i++, off += sizeof sheader) { read_len = load_fread(&sheader, 1, sizeof sheader, fd); check_read(read_len, sizeof sheader); if (strcmp(p_section_str + sheader.sh_name, ".got") == 0) { got_start = (void *)((uint8_t *)sheader.sh_addr + load_off); got_size = (size_t)sheader.sh_size; } else if (strcmp(p_section_str + sheader.sh_name, ".rel.dyn") == 0) { rel_dyn_start = (void *)((uint8_t *)sheader.sh_addr + load_off); rel_dyn_size = (size_t)sheader.sh_size; } else if (strcmp(p_section_str + sheader.sh_name, ".dynsym") == 0) { dynsym_off = (size_t)sheader.sh_offset; dynsym_size = (size_t)sheader.sh_size; } } /* reloc */ if (dynsym_size) { dynsym = rt_malloc(dynsym_size); if (!dynsym) { LOG_E("ERROR: Malloc error!"); result = -ENOMEM; goto _exit; } check_off(dynsym_off, len); lseek(fd, dynsym_off, SEEK_SET); read_len = load_fread(dynsym, 1, dynsym_size, fd); check_read(read_len, dynsym_size); } #ifdef ARCH_MM_MMU arch_elf_reloc(lwp->aspace, (void *)load_off, rel_dyn_start, rel_dyn_size, got_start, got_size, dynsym); #else arch_elf_reloc((void *)load_off, rel_dyn_start, rel_dyn_size, got_start, got_size, dynsym); rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, lwp->text_entry, lwp->text_size); rt_hw_cpu_icache_ops(RT_HW_CACHE_INVALIDATE, lwp->text_entry, lwp->text_size); #endif } LOG_D("lwp->text_entry = 0x%p", lwp->text_entry); LOG_D("lwp->text_size = 0x%p", lwp->text_size); _exit: if (dynsym) { rt_free(dynsym); } if (p_section_str) { rt_free(p_section_str); } if (result != RT_EOK) { LOG_E("lwp load faild, %d", result); } return result; } #endif /* ARCH_MM_MMU */ rt_weak int lwp_load(const char *filename, struct rt_lwp *lwp, uint8_t *load_addr, size_t addr_size, struct process_aux *aux) { uint8_t *ptr; int ret = -1; int len; int fd = -1; /* check file name */ RT_ASSERT(filename != RT_NULL); /* check lwp control block */ RT_ASSERT(lwp != RT_NULL); /* copy file name to process name */ rt_strncpy(lwp->cmd, filename, RT_NAME_MAX); if (load_addr != RT_NULL) { lwp->lwp_type = LWP_TYPE_FIX_ADDR; ptr = load_addr; } else { lwp->lwp_type = LWP_TYPE_DYN_ADDR; ptr = RT_NULL; } fd = open(filename, O_BINARY | O_RDONLY, 0); if (fd < 0) { LOG_E("ERROR: Can't open elf file %s!", filename); goto out; } len = lseek(fd, 0, SEEK_END); if (len < 0) { LOG_E("ERROR: File %s size error!", filename); goto out; } lseek(fd, 0, SEEK_SET); ret = load_elf(fd, len, lwp, ptr, aux); if ((ret != RT_EOK) && (ret != 1)) { LOG_E("lwp load ret = %d", ret); } out: if (fd > 0) { close(fd); } return ret; } /* lwp thread clean up */ void lwp_cleanup(struct rt_thread *tid) { rt_base_t level; struct rt_lwp *lwp; if (tid == NULL) { LOG_I("%s: invalid parameter tid == NULL", __func__); return; } else LOG_D("cleanup thread: %s, stack_addr: 0x%x", tid->parent.name, tid->stack_addr); level = rt_hw_interrupt_disable(); lwp = (struct rt_lwp *)tid->lwp; /* lwp thread cleanup */ lwp_tid_put(tid->tid); rt_list_remove(&tid->sibling); lwp_thread_signal_detach(&tid->signal); rt_hw_interrupt_enable(level); /* tty will be release in lwp_ref_dec() if ref is cleared */ lwp_ref_dec(lwp); return; } static void lwp_copy_stdio_fdt(struct rt_lwp *lwp) { struct dfs_file *d; struct dfs_fdtable *lwp_fdt; lwp_fdt = &lwp->fdt; /* init 4 fds */ lwp_fdt->fds = rt_calloc(4, sizeof(void *)); if (lwp_fdt->fds) { lwp_fdt->maxfd = 4; d = fd_get(0); fd_associate(lwp_fdt, 0, d); d = fd_get(1); fd_associate(lwp_fdt, 1, d); d = fd_get(2); fd_associate(lwp_fdt, 2, d); } return; } static void _lwp_thread_entry(void *parameter) { rt_thread_t tid; struct rt_lwp *lwp; tid = rt_thread_self(); lwp = (struct rt_lwp *)tid->lwp; tid->cleanup = lwp_cleanup; tid->user_stack = RT_NULL; if (lwp->debug) { lwp->bak_first_ins = *(uint32_t *)lwp->text_entry; *(uint32_t *)lwp->text_entry = dbg_get_ins(); rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, lwp->text_entry, sizeof(uint32_t)); icache_invalid_all(); } /** * without ASID support, it will be a special case when trying to run application * and exit multiple times and a same page frame allocated to it bound to * different text segment. Then we are in a situation where icache contains * out-of-dated data and must be handle by the running core itself. * with ASID support, this should be a rare case that ASID & page frame both * identical to previous running application. * * For a new application loaded into memory, icache are seen as empty. And there * should be nothing in the icache entry to match. So this icache invalidation * operation should have barely influence. */ rt_hw_icache_invalidate_all(); #ifdef ARCH_MM_MMU arch_start_umode(lwp->args, lwp->text_entry, (void *)USER_STACK_VEND, (char *)tid->stack_addr + tid->stack_size); #else arch_start_umode(lwp->args, lwp->text_entry, lwp->data_entry, (void *)((uint32_t)lwp->data_entry + lwp->data_size)); #endif /* ARCH_MM_MMU */ } struct rt_lwp *lwp_self(void) { rt_thread_t tid; tid = rt_thread_self(); if (tid) { return (struct rt_lwp *)tid->lwp; } return RT_NULL; } pid_t lwp_execve(char *filename, int debug, int argc, char **argv, char **envp) { int result; rt_base_t level; struct rt_lwp *lwp; char *thread_name; char *argv_last = argv[argc - 1]; int bg = 0; struct process_aux *aux; int tid = 0; int ret; if (filename == RT_NULL) { return -RT_ERROR; } if (access(filename, X_OK) != 0) { return -EACCES; } lwp = lwp_new(); if (lwp == RT_NULL) { dbg_log(DBG_ERROR, "lwp struct out of memory!\n"); return -RT_ENOMEM; } LOG_D("lwp malloc : %p, size: %d!", lwp, sizeof(struct rt_lwp)); if ((tid = lwp_tid_get()) == 0) { lwp_ref_dec(lwp); return -ENOMEM; } #ifdef ARCH_MM_MMU if (lwp_user_space_init(lwp, 0) != 0) { lwp_tid_put(tid); lwp_ref_dec(lwp); return -ENOMEM; } #endif if (argv_last[0] == '&' && argv_last[1] == '\0') { argc--; bg = 1; } if ((aux = lwp_argscopy(lwp, argc, argv, envp)) == RT_NULL) { lwp_tid_put(tid); lwp_ref_dec(lwp); return -ENOMEM; } result = lwp_load(filename, lwp, RT_NULL, 0, aux); #ifdef ARCH_MM_MMU if (result == 1) { /* dynmaic */ lwp_unmap_user(lwp, (void *)(USER_VADDR_TOP - ARCH_PAGE_SIZE)); result = load_ldso(lwp, filename, argv, envp); } #endif /* ARCH_MM_MMU */ if (result == RT_EOK) { rt_thread_t thread = RT_NULL; rt_uint32_t priority = 25, tick = 200; lwp_copy_stdio_fdt(lwp); /* obtain the base name */ thread_name = strrchr(filename, '/'); thread_name = thread_name ? thread_name + 1 : filename; #ifndef ARCH_MM_MMU struct lwp_app_head *app_head = lwp->text_entry; if (app_head->priority) { priority = app_head->priority; } if (app_head->tick) { tick = app_head->tick; } #endif /* not defined ARCH_MM_MMU */ thread = rt_thread_create(thread_name, _lwp_thread_entry, RT_NULL, LWP_TASK_STACK_SIZE, priority, tick); if (thread != RT_NULL) { struct rt_lwp *self_lwp; thread->tid = tid; lwp_tid_set_thread(tid, thread); LOG_D("lwp kernel => (0x%08x, 0x%08x)\n", (rt_size_t)thread->stack_addr, (rt_size_t)thread->stack_addr + thread->stack_size); level = rt_hw_interrupt_disable(); self_lwp = lwp_self(); if (self_lwp) { //lwp->tgroup_leader = &thread; //add thread group leader for lwp lwp->__pgrp = tid; lwp->session = self_lwp->session; /* lwp add to children link */ lwp->sibling = self_lwp->first_child; self_lwp->first_child = lwp; lwp->parent = self_lwp; } else { //lwp->tgroup_leader = &thread; //add thread group leader for lwp lwp->__pgrp = tid; } if (!bg) { if (lwp->session == -1) { struct tty_struct *tty = RT_NULL; struct rt_lwp *old_lwp; tty = (struct tty_struct *)console_tty_get(); old_lwp = tty->foreground; if (old_lwp) { rt_mutex_take(&tty->lock, RT_WAITING_FOREVER); ret = tty_push(&tty->head, old_lwp); rt_mutex_release(&tty->lock); if (ret < 0) { lwp_tid_put(tid); lwp_ref_dec(lwp); LOG_E("malloc fail!\n"); return -ENOMEM; } } lwp->tty = tty; lwp->tty->pgrp = lwp->__pgrp; lwp->tty->session = lwp->session; lwp->tty->foreground = lwp; tcgetattr(1, &stdin_termios); old_stdin_termios = stdin_termios; stdin_termios.c_lflag |= ICANON | ECHO | ECHOCTL; tcsetattr(1, 0, &stdin_termios); } else { if (self_lwp != RT_NULL) { rt_mutex_take(&self_lwp->tty->lock, RT_WAITING_FOREVER); ret = tty_push(&self_lwp->tty->head, self_lwp); rt_mutex_release(&self_lwp->tty->lock); if (ret < 0) { lwp_tid_put(tid); lwp_ref_dec(lwp); LOG_E("malloc fail!\n"); return -ENOMEM; } lwp->tty = self_lwp->tty; lwp->tty->pgrp = lwp->__pgrp; lwp->tty->session = lwp->session; lwp->tty->foreground = lwp; } else { lwp->tty = RT_NULL; } } } else { lwp->background = RT_TRUE; } thread->lwp = lwp; #ifndef ARCH_MM_MMU struct lwp_app_head *app_head = (struct lwp_app_head*)lwp->text_entry; thread->user_stack = app_head->stack_offset ? (void *)(app_head->stack_offset - app_head->data_offset + (uint32_t)lwp->data_entry) : RT_NULL; thread->user_stack_size = app_head->stack_size; /* init data area */ rt_memset(lwp->data_entry, 0, lwp->data_size); /* init user stack */ rt_memset(thread->user_stack, '#', thread->user_stack_size); #endif /* not defined ARCH_MM_MMU */ rt_list_insert_after(&lwp->t_grp, &thread->sibling); if (debug && rt_dbg_ops) { lwp->debug = debug; rt_thread_control(thread, RT_THREAD_CTRL_BIND_CPU, (void*)0); } rt_hw_interrupt_enable(level); rt_thread_startup(thread); return lwp_to_pid(lwp); } } lwp_tid_put(tid); lwp_ref_dec(lwp); return -RT_ERROR; } #ifdef RT_USING_MUSL extern char **__environ; #else char **__environ = 0; #endif pid_t exec(char *filename, int debug, int argc, char **argv) { setenv("OS", "RT-Thread", 1); return lwp_execve(filename, debug, argc, argv, __environ); } #ifdef ARCH_MM_MMU void lwp_user_setting_save(rt_thread_t thread) { if (thread) { thread->thread_idr = arch_get_tidr(); } } void lwp_user_setting_restore(rt_thread_t thread) { if (!thread) { return; } #if !defined(ARCH_RISCV64) /* tidr will be set in RESTORE_ALL in risc-v */ arch_set_tidr(thread->thread_idr); #endif if (rt_dbg_ops) { struct rt_lwp *l = (struct rt_lwp *)thread->lwp; if (l != 0) { rt_hw_set_process_id((size_t)l->pid); } else { rt_hw_set_process_id(0); } if (l && l->debug) { uint32_t step_type = 0; step_type = dbg_step_type(); if ((step_type == 2) || (thread->step_exec && (step_type == 1))) { dbg_activate_step(); } else { dbg_deactivate_step(); } } } } #endif /* ARCH_MM_MMU */