/* * Copyright (c) 2006-2020, RT-Thread Development Team * * SPDX-License-Identifier: Apache-2.0 * * Change Logs: * Date Author Notes * 2019-10-12 Jesven first version * 2023-02-20 wangxiaoyao adapt to mm */ #include #include #ifdef ARCH_MM_MMU #include #include #include #include #include /* the kernel structure to represent a share-memory */ struct lwp_shm_struct { struct rt_mem_obj mem_obj; size_t addr; /* point to the next item in the free list when not used */ size_t size; int ref; size_t key; }; static struct lwp_avl_struct *shm_tree_key; static struct lwp_avl_struct *shm_tree_pa; static int shm_free_list = -1; /* the single-direct list of freed items */ static int shm_id_used = 0; /* the latest allocated item in the array */ static struct lwp_shm_struct _shm_ary[RT_LWP_SHM_MAX_NR]; static const char *get_shm_name(rt_varea_t varea) { return "user.shm"; } static void on_shm_varea_open(struct rt_varea *varea) { struct lwp_shm_struct *shm; shm = rt_container_of(varea->mem_obj, struct lwp_shm_struct, mem_obj); shm->ref += 1; } static void on_shm_varea_close(struct rt_varea *varea) { struct lwp_shm_struct *shm; shm = rt_container_of(varea->mem_obj, struct lwp_shm_struct, mem_obj); shm->ref -= 1; } static void on_shm_page_fault(struct rt_varea *varea, struct rt_aspace_fault_msg *msg) { struct lwp_shm_struct *shm; int err; shm = rt_container_of(varea->mem_obj, struct lwp_shm_struct, mem_obj); /* map all share page frames to user space in a time */ void *page = (void *)shm->addr; void *pg_paddr = (char *)page + PV_OFFSET; err = rt_varea_map_range(varea, varea->start, pg_paddr, shm->size); if (err == RT_EOK) { msg->response.status = MM_FAULT_STATUS_OK_MAPPED; msg->response.size = shm->size; msg->response.vaddr = page; } return ; } /* * Try to allocate an structure 'lwp_shm_struct' from the freed list or the * static array. */ static int _shm_id_alloc(void) { int id = -1; if (shm_free_list != -1) /* first try the freed list */ { id = shm_free_list; shm_free_list = (int)_shm_ary[shm_free_list].addr; /* single-direction */ } else if (shm_id_used < RT_LWP_SHM_MAX_NR) /* then try the array */ { id = shm_id_used; shm_id_used++; } return id; } /* Release the item in the static array to the freed list. */ static void shm_id_free(int id) { /* link the freed itme to the single-direction list */ _shm_ary[id].addr = (size_t)shm_free_list; shm_free_list = id; } /* Locate the shared memory through 'key' or create a new one. */ static int _lwp_shmget(size_t key, size_t size, int create) { int id = -1; struct lwp_avl_struct *node_key = 0; struct lwp_avl_struct *node_pa = 0; void *page_addr = 0; uint32_t bit = 0; /* try to locate the item with the key in the binary tree */ node_key = lwp_avl_find(key, shm_tree_key); if (node_key) { return (struct lwp_shm_struct *)node_key->data - _shm_ary; /* the index */ } /* If there doesn't exist such an item and we're allowed to create one ... */ if (create) { struct lwp_shm_struct* p; if (!size) { goto err; } id = _shm_id_alloc(); if (id == -1) { goto err; } /* allocate pages up to 2's exponent to cover the required size */ bit = rt_page_bits(size); page_addr = rt_pages_alloc_ext(bit, PAGE_ANY_AVAILABLE); /* virtual address */ if (!page_addr) { goto err; } /* initialize the shared memory structure */ p = _shm_ary + id; p->addr = (size_t)page_addr; p->size = (1UL << (bit + ARCH_PAGE_SHIFT)); p->ref = 0; p->key = key; p->mem_obj.get_name = get_shm_name; p->mem_obj.on_page_fault = on_shm_page_fault; p->mem_obj.on_varea_open = on_shm_varea_open; p->mem_obj.on_varea_close = on_shm_varea_close; p->mem_obj.hint_free = NULL; p->mem_obj.on_page_offload = NULL; /* then insert it into the balancing binary tree */ node_key = (struct lwp_avl_struct *)rt_malloc(sizeof(struct lwp_avl_struct) * 2); if (!node_key) { goto err; } node_key->avl_key = p->key; node_key->data = (void *)p; lwp_avl_insert(node_key, &shm_tree_key); node_pa = node_key + 1; node_pa->avl_key = p->addr; node_pa->data = (void *)p; lwp_avl_insert(node_pa, &shm_tree_pa); } return id; err: if (id != -1) { shm_id_free(id); } if (page_addr) { rt_pages_free(page_addr, bit); } if (node_key) { rt_free(node_key); } return -1; } /* A wrapping function, get the shared memory with interrupts disabled. */ int lwp_shmget(size_t key, size_t size, int create) { int ret = 0; rt_mm_lock(); ret = _lwp_shmget(key, size, create); rt_mm_unlock(); return ret; } /* Locate the binary tree node_key corresponding to the shared-memory id. */ static struct lwp_avl_struct *shm_id_to_node(int id) { struct lwp_avl_struct *node_key = 0; struct lwp_shm_struct *p = RT_NULL; /* check id */ if (id < 0 || id >= RT_LWP_SHM_MAX_NR) { return RT_NULL; } p = _shm_ary + id; /* the address of the shared-memory structure */ node_key = lwp_avl_find(p->key, shm_tree_key); if (!node_key) { return RT_NULL; } if (node_key->data != (void *)p) { return RT_NULL; } return node_key; } /* Free the shared pages, the shared-memory structure and its binary tree node_key. */ static int _lwp_shmrm(int id) { struct lwp_avl_struct *node_key = RT_NULL; struct lwp_avl_struct *node_pa = RT_NULL; struct lwp_shm_struct* p = RT_NULL; uint32_t bit = 0; node_key = shm_id_to_node(id); if (!node_key) { return -1; } p = (struct lwp_shm_struct *)node_key->data; if (p->ref) { return 0; } bit = rt_page_bits(p->size); rt_pages_free((void *)p->addr, bit); lwp_avl_remove(node_key, &shm_tree_key); node_pa = node_key + 1; lwp_avl_remove(node_pa, &shm_tree_pa); rt_free(node_key); shm_id_free(id); return 0; } /* A wrapping function, free the shared memory with interrupt disabled. */ int lwp_shmrm(int id) { int ret = 0; ret = _lwp_shmrm(id); return ret; } /* Map the shared memory specified by 'id' to the specified virtual address. */ static void *_lwp_shmat(int id, void *shm_vaddr) { int err; struct rt_lwp *lwp = RT_NULL; struct lwp_avl_struct *node_key = RT_NULL; struct lwp_shm_struct *p = RT_NULL; void *va = shm_vaddr; /* The id is used to locate the node_key in the binary tree, and then get the * shared-memory structure linked to the node_key. We don't use the id to refer * to the shared-memory structure directly, because the binary tree is used * to verify the structure is really in use. */ node_key = shm_id_to_node(id); if (!node_key) { return RT_NULL; } p = (struct lwp_shm_struct *)node_key->data; /* p = _shm_ary[id]; */ /* map the shared memory into the address space of the current thread */ lwp = lwp_self(); if (!lwp) { return RT_NULL; } err = rt_aspace_map(lwp->aspace, &va, p->size, MMU_MAP_U_RWCB, MMF_PREFETCH, &p->mem_obj, 0); if (err != RT_EOK) { va = RT_NULL; } return va; } /* A wrapping function: attach the shared memory to the specified address. */ void *lwp_shmat(int id, void *shm_vaddr) { void *ret = RT_NULL; if (((size_t)shm_vaddr & ARCH_PAGE_MASK) != 0) { return RT_NULL; } ret = _lwp_shmat(id, shm_vaddr); return ret; } static struct lwp_shm_struct *_lwp_shm_struct_get(struct rt_lwp *lwp, void *shm_vaddr) { void *pa = RT_NULL; struct lwp_avl_struct *node_pa = RT_NULL; if (!lwp) { return RT_NULL; } pa = lwp_v2p(lwp, shm_vaddr); /* physical memory */ node_pa = lwp_avl_find((size_t)pa, shm_tree_pa); if (!node_pa) { return RT_NULL; } return (struct lwp_shm_struct *)node_pa->data; } static int _lwp_shm_ref_inc(struct rt_lwp *lwp, void *shm_vaddr) { struct lwp_shm_struct* p = _lwp_shm_struct_get(lwp, shm_vaddr); if (p) { p->ref++; return p->ref; } return -1; } int lwp_shm_ref_inc(struct rt_lwp *lwp, void *shm_vaddr) { int ret = 0; rt_mm_lock(); ret = _lwp_shm_ref_inc(lwp, shm_vaddr); rt_mm_unlock(); return ret; } static int _lwp_shm_ref_dec(struct rt_lwp *lwp, void *shm_vaddr) { struct lwp_shm_struct* p = _lwp_shm_struct_get(lwp, shm_vaddr); if (p && (p->ref > 0)) { p->ref--; return p->ref; } return -1; } int lwp_shm_ref_dec(struct rt_lwp *lwp, void *shm_vaddr) { int ret = 0; rt_mm_lock(); ret = _lwp_shm_ref_dec(lwp, shm_vaddr); rt_mm_unlock(); return ret; } /* Unmap the shared memory from the address space of the current thread. */ int _lwp_shmdt(void *shm_vaddr) { struct rt_lwp *lwp = RT_NULL; int ret = 0; lwp = lwp_self(); if (!lwp) { return -1; } ret = rt_aspace_unmap(lwp->aspace, shm_vaddr); if (ret != RT_EOK) { ret = -1; } return ret; } /* A wrapping function: detach the mapped shared memory. */ int lwp_shmdt(void *shm_vaddr) { int ret = 0; rt_mm_lock(); ret = _lwp_shmdt(shm_vaddr); rt_mm_unlock(); return ret; } /* Get the virtual address of a shared memory in kernel. */ void *_lwp_shminfo(int id) { struct lwp_avl_struct *node_key = RT_NULL; struct lwp_shm_struct *p = RT_NULL; /* the share memory is in use only if it exsits in the binary tree */ node_key = shm_id_to_node(id); if (!node_key) { return RT_NULL; } p = (struct lwp_shm_struct *)node_key->data; /* p = _shm_ary[id]; */ return (void *)((char *)p->addr - PV_OFFSET); /* get the virtual address */ } /* A wrapping function: get the virtual address of a shared memory. */ void *lwp_shminfo(int id) { void *vaddr = RT_NULL; rt_mm_lock(); vaddr = _lwp_shminfo(id); rt_mm_unlock(); return vaddr; } #ifdef RT_USING_FINSH static int _shm_info(struct lwp_avl_struct* node_key, void *data) { int id = 0; struct lwp_shm_struct* p = (struct lwp_shm_struct *)node_key->data; id = p - _shm_ary; rt_kprintf("0x%08x 0x%08x 0x%08x %8d\n", p->key, p->addr, p->size, id); return 0; } void list_shm(void) { rt_kprintf(" key paddr size id\n"); rt_kprintf("---------- ---------- ---------- --------\n"); rt_mm_lock(); lwp_avl_traversal(shm_tree_key, _shm_info, NULL); rt_mm_unlock(); } MSH_CMD_EXPORT(list_shm, show share memory info); #endif #endif