2022-12-03 12:07:44 +08:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2006-2020, RT-Thread Development Team
|
|
|
|
*
|
|
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
|
|
*
|
|
|
|
* Change Logs:
|
|
|
|
* Date Author Notes
|
|
|
|
* 2019-10-12 Jesven first version
|
|
|
|
*/
|
|
|
|
#include <rthw.h>
|
|
|
|
#include <rtthread.h>
|
|
|
|
|
2022-12-16 18:38:28 +08:00
|
|
|
#ifdef ARCH_MM_MMU
|
2022-12-03 12:07:44 +08:00
|
|
|
#include <lwp.h>
|
|
|
|
#include <lwp_shm.h>
|
|
|
|
#include <lwp_mm.h>
|
|
|
|
|
|
|
|
#include <lwp_mm_area.h>
|
|
|
|
#include <lwp_user_mm.h>
|
|
|
|
|
|
|
|
/* the kernel structure to represent a share-memory */
|
|
|
|
struct lwp_shm_struct
|
|
|
|
{
|
|
|
|
size_t addr; /* point to the next item in the free list when not used */
|
|
|
|
size_t size;
|
|
|
|
int ref;
|
|
|
|
size_t key;
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct lwp_avl_struct *shm_tree_key;
|
|
|
|
static struct lwp_avl_struct *shm_tree_pa;
|
|
|
|
|
|
|
|
static int shm_free_list = -1; /* the single-direct list of freed items */
|
|
|
|
static int shm_id_used = 0; /* the latest allocated item in the array */
|
|
|
|
static struct lwp_shm_struct _shm_ary[RT_LWP_SHM_MAX_NR];
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Try to allocate an structure 'lwp_shm_struct' from the freed list or the
|
|
|
|
* static array.
|
|
|
|
*/
|
|
|
|
static int _shm_id_alloc(void)
|
|
|
|
{
|
|
|
|
int id = -1;
|
|
|
|
|
|
|
|
if (shm_free_list != -1) /* first try the freed list */
|
|
|
|
{
|
|
|
|
id = shm_free_list;
|
|
|
|
shm_free_list = (int)_shm_ary[shm_free_list].addr; /* single-direction */
|
|
|
|
}
|
|
|
|
else if (shm_id_used < RT_LWP_SHM_MAX_NR) /* then try the array */
|
|
|
|
{
|
|
|
|
id = shm_id_used;
|
|
|
|
shm_id_used++;
|
|
|
|
}
|
|
|
|
return id;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Release the item in the static array to the freed list. */
|
|
|
|
static void shm_id_free(int id)
|
|
|
|
{
|
|
|
|
/* link the freed itme to the single-direction list */
|
|
|
|
_shm_ary[id].addr = (size_t)shm_free_list;
|
|
|
|
shm_free_list = id;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Locate the shared memory through 'key' or create a new one. */
|
|
|
|
static int _lwp_shmget(size_t key, size_t size, int create)
|
|
|
|
{
|
|
|
|
int id = -1;
|
|
|
|
struct lwp_avl_struct *node_key = 0;
|
|
|
|
struct lwp_avl_struct *node_pa = 0;
|
|
|
|
void *page_addr = 0, *page_addr_p = RT_NULL;
|
|
|
|
uint32_t bit = 0;
|
|
|
|
|
|
|
|
/* try to locate the item with the key in the binary tree */
|
|
|
|
node_key = lwp_avl_find(key, shm_tree_key);
|
|
|
|
if (node_key)
|
|
|
|
{
|
|
|
|
return (struct lwp_shm_struct *)node_key->data - _shm_ary; /* the index */
|
|
|
|
}
|
|
|
|
|
|
|
|
/* If there doesn't exist such an item and we're allowed to create one ... */
|
|
|
|
if (create)
|
|
|
|
{
|
|
|
|
struct lwp_shm_struct* p;
|
|
|
|
|
|
|
|
if (!size)
|
|
|
|
{
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
id = _shm_id_alloc();
|
|
|
|
if (id == -1)
|
|
|
|
{
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* allocate pages up to 2's exponent to cover the required size */
|
|
|
|
bit = rt_page_bits(size);
|
|
|
|
page_addr = rt_pages_alloc(bit); /* virtual address */
|
|
|
|
if (!page_addr)
|
|
|
|
{
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
page_addr_p = (void *)((char *)page_addr + PV_OFFSET); /* physical address */
|
|
|
|
|
|
|
|
/* initialize the shared memory structure */
|
|
|
|
p = _shm_ary + id;
|
|
|
|
p->addr = (size_t)page_addr_p;
|
|
|
|
p->size = (1UL << (bit + ARCH_PAGE_SHIFT));
|
|
|
|
p->ref = 0;
|
|
|
|
p->key = key;
|
|
|
|
|
|
|
|
/* then insert it into the balancing binary tree */
|
|
|
|
node_key = (struct lwp_avl_struct *)rt_malloc(sizeof(struct lwp_avl_struct) * 2);
|
|
|
|
if (!node_key)
|
|
|
|
{
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
node_key->avl_key = p->key;
|
|
|
|
node_key->data = (void *)p;
|
|
|
|
lwp_avl_insert(node_key, &shm_tree_key);
|
|
|
|
node_pa = node_key + 1;
|
|
|
|
node_pa->avl_key = p->addr;
|
|
|
|
node_pa->data = (void *)p;
|
|
|
|
lwp_avl_insert(node_pa, &shm_tree_pa);
|
|
|
|
}
|
|
|
|
return id;
|
|
|
|
|
|
|
|
err:
|
|
|
|
if (id != -1)
|
|
|
|
{
|
|
|
|
shm_id_free(id);
|
|
|
|
}
|
|
|
|
if (page_addr)
|
|
|
|
{
|
|
|
|
rt_pages_free(page_addr, bit);
|
|
|
|
}
|
|
|
|
if (node_key)
|
|
|
|
{
|
|
|
|
rt_free(node_key);
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* A wrapping function, get the shared memory with interrupts disabled. */
|
|
|
|
int lwp_shmget(size_t key, size_t size, int create)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
rt_mm_lock();
|
|
|
|
ret = _lwp_shmget(key, size, create);
|
|
|
|
rt_mm_unlock();
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Locate the binary tree node_key corresponding to the shared-memory id. */
|
|
|
|
static struct lwp_avl_struct *shm_id_to_node(int id)
|
|
|
|
{
|
|
|
|
struct lwp_avl_struct *node_key = 0;
|
|
|
|
struct lwp_shm_struct *p = RT_NULL;
|
|
|
|
|
|
|
|
/* check id */
|
|
|
|
if (id < 0 || id >= RT_LWP_SHM_MAX_NR)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
p = _shm_ary + id; /* the address of the shared-memory structure */
|
|
|
|
node_key = lwp_avl_find(p->key, shm_tree_key);
|
|
|
|
if (!node_key)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
if (node_key->data != (void *)p)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
return node_key;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Free the shared pages, the shared-memory structure and its binary tree node_key. */
|
|
|
|
static int _lwp_shmrm(int id)
|
|
|
|
{
|
|
|
|
struct lwp_avl_struct *node_key = RT_NULL;
|
|
|
|
struct lwp_avl_struct *node_pa = RT_NULL;
|
|
|
|
struct lwp_shm_struct* p = RT_NULL;
|
|
|
|
uint32_t bit = 0;
|
|
|
|
|
|
|
|
node_key = shm_id_to_node(id);
|
|
|
|
if (!node_key)
|
|
|
|
{
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
p = (struct lwp_shm_struct *)node_key->data;
|
|
|
|
if (p->ref)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
bit = rt_page_bits(p->size);
|
|
|
|
rt_pages_free((void *)((char *)p->addr - PV_OFFSET), bit);
|
|
|
|
lwp_avl_remove(node_key, &shm_tree_key);
|
|
|
|
node_pa = node_key + 1;
|
|
|
|
lwp_avl_remove(node_pa, &shm_tree_pa);
|
|
|
|
rt_free(node_key);
|
|
|
|
shm_id_free(id);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* A wrapping function, free the shared memory with interrupt disabled. */
|
|
|
|
int lwp_shmrm(int id)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
rt_mm_lock();
|
|
|
|
ret = _lwp_shmrm(id);
|
|
|
|
rt_mm_unlock();
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Map the shared memory specified by 'id' to the specified virtual address. */
|
|
|
|
static void *_lwp_shmat(int id, void *shm_vaddr)
|
|
|
|
{
|
|
|
|
struct rt_lwp *lwp = RT_NULL;
|
|
|
|
struct lwp_avl_struct *node_key = RT_NULL;
|
|
|
|
struct lwp_shm_struct *p = RT_NULL;
|
|
|
|
void *va = RT_NULL;
|
|
|
|
|
|
|
|
/* The id is used to locate the node_key in the binary tree, and then get the
|
|
|
|
* shared-memory structure linked to the node_key. We don't use the id to refer
|
|
|
|
* to the shared-memory structure directly, because the binary tree is used
|
|
|
|
* to verify the structure is really in use.
|
|
|
|
*/
|
|
|
|
node_key = shm_id_to_node(id);
|
|
|
|
if (!node_key)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
p = (struct lwp_shm_struct *)node_key->data; /* p = _shm_ary[id]; */
|
|
|
|
|
|
|
|
/* map the shared memory into the address space of the current thread */
|
|
|
|
lwp = lwp_self();
|
|
|
|
if (!lwp)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
va = lwp_map_user_type(lwp, shm_vaddr, (void *)p->addr, p->size, 1, MM_AREA_TYPE_SHM);
|
|
|
|
if (va)
|
|
|
|
{
|
|
|
|
p->ref++;
|
|
|
|
}
|
|
|
|
return va;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* A wrapping function: attach the shared memory to the specified address. */
|
|
|
|
void *lwp_shmat(int id, void *shm_vaddr)
|
|
|
|
{
|
|
|
|
void *ret = RT_NULL;
|
|
|
|
|
|
|
|
if (((size_t)shm_vaddr & ARCH_PAGE_MASK) != 0)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
rt_mm_lock();
|
|
|
|
ret = _lwp_shmat(id, shm_vaddr);
|
|
|
|
rt_mm_unlock();
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct lwp_shm_struct *_lwp_shm_struct_get(struct rt_lwp *lwp, void *shm_vaddr)
|
|
|
|
{
|
|
|
|
void *pa = RT_NULL;
|
|
|
|
struct lwp_avl_struct *node_pa = RT_NULL;
|
|
|
|
|
|
|
|
if (!lwp)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
pa = rt_hw_mmu_v2p(&lwp->mmu_info, shm_vaddr); /* physical memory */
|
|
|
|
|
|
|
|
node_pa = lwp_avl_find((size_t)pa, shm_tree_pa);
|
|
|
|
if (!node_pa)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
return (struct lwp_shm_struct *)node_pa->data;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int _lwp_shm_ref_inc(struct rt_lwp *lwp, void *shm_vaddr)
|
|
|
|
{
|
|
|
|
struct lwp_shm_struct* p = _lwp_shm_struct_get(lwp, shm_vaddr);
|
|
|
|
|
|
|
|
if (p)
|
|
|
|
{
|
|
|
|
p->ref++;
|
|
|
|
return p->ref;
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
int lwp_shm_ref_inc(struct rt_lwp *lwp, void *shm_vaddr)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
rt_mm_lock();
|
|
|
|
ret = _lwp_shm_ref_inc(lwp, shm_vaddr);
|
|
|
|
rt_mm_unlock();
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int _lwp_shm_ref_dec(struct rt_lwp *lwp, void *shm_vaddr)
|
|
|
|
{
|
|
|
|
struct lwp_shm_struct* p = _lwp_shm_struct_get(lwp, shm_vaddr);
|
|
|
|
|
|
|
|
if (p && (p->ref > 0))
|
|
|
|
{
|
|
|
|
p->ref--;
|
|
|
|
return p->ref;
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
int lwp_shm_ref_dec(struct rt_lwp *lwp, void *shm_vaddr)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
rt_mm_lock();
|
|
|
|
ret = _lwp_shm_ref_dec(lwp, shm_vaddr);
|
|
|
|
rt_mm_unlock();
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Unmap the shared memory from the address space of the current thread. */
|
|
|
|
int _lwp_shmdt(void *shm_vaddr)
|
|
|
|
{
|
|
|
|
struct rt_lwp *lwp = RT_NULL;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
lwp = lwp_self();
|
|
|
|
if (!lwp)
|
|
|
|
{
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
ret = _lwp_shm_ref_dec(lwp, shm_vaddr);
|
|
|
|
if (ret >= 0)
|
|
|
|
{
|
|
|
|
lwp_unmap_user_phy(lwp, shm_vaddr);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
/* A wrapping function: detach the mapped shared memory. */
|
|
|
|
int lwp_shmdt(void *shm_vaddr)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
rt_mm_lock();
|
|
|
|
ret = _lwp_shmdt(shm_vaddr);
|
|
|
|
rt_mm_unlock();
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Get the virtual address of a shared memory in kernel. */
|
|
|
|
void *_lwp_shminfo(int id)
|
|
|
|
{
|
|
|
|
struct lwp_avl_struct *node_key = RT_NULL;
|
|
|
|
struct lwp_shm_struct *p = RT_NULL;
|
|
|
|
|
|
|
|
/* the share memory is in use only if it exsits in the binary tree */
|
|
|
|
node_key = shm_id_to_node(id);
|
|
|
|
if (!node_key)
|
|
|
|
{
|
|
|
|
return RT_NULL;
|
|
|
|
}
|
|
|
|
p = (struct lwp_shm_struct *)node_key->data; /* p = _shm_ary[id]; */
|
|
|
|
|
|
|
|
return (void *)((char *)p->addr - PV_OFFSET); /* get the virtual address */
|
|
|
|
}
|
|
|
|
|
|
|
|
/* A wrapping function: get the virtual address of a shared memory. */
|
|
|
|
void *lwp_shminfo(int id)
|
|
|
|
{
|
|
|
|
void *vaddr = RT_NULL;
|
|
|
|
|
|
|
|
rt_mm_lock();
|
|
|
|
vaddr = _lwp_shminfo(id);
|
|
|
|
rt_mm_unlock();
|
|
|
|
return vaddr;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef RT_USING_FINSH
|
|
|
|
static int _shm_info(struct lwp_avl_struct* node_key, void *data)
|
|
|
|
{
|
|
|
|
int id = 0;
|
|
|
|
struct lwp_shm_struct* p = (struct lwp_shm_struct *)node_key->data;
|
|
|
|
|
|
|
|
id = p - _shm_ary;
|
|
|
|
rt_kprintf("0x%08x 0x%08x 0x%08x %8d\n", p->key, p->addr, p->size, id);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void list_shm(void)
|
|
|
|
{
|
|
|
|
rt_kprintf(" key paddr size id\n");
|
|
|
|
rt_kprintf("---------- ---------- ---------- --------\n");
|
|
|
|
rt_mm_lock();
|
|
|
|
lwp_avl_traversal(shm_tree_key, _shm_info, NULL);
|
|
|
|
rt_mm_unlock();
|
|
|
|
}
|
|
|
|
MSH_CMD_EXPORT(list_shm, show share memory info);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif
|