/* * Copyright (c) 2006-2023, RT-Thread Development Team * * SPDX-License-Identifier: Apache-2.0 * * Change Logs: * Date Author Notes * 2023-02-25 GuEe-GUI the first version */ #include #include #include #define DBG_TAG "dma.pool" #define DBG_LVL DBG_INFO #include #include #include static struct rt_spinlock dma_pools_lock = {}; static rt_list_t dma_pool_nodes = RT_LIST_OBJECT_INIT(dma_pool_nodes); static struct rt_dma_pool *dma_pool_install(rt_region_t *region); static void *dma_alloc(struct rt_device *dev, rt_size_t size, rt_ubase_t *dma_handle, rt_ubase_t flags); static void dma_free(struct rt_device *dev, rt_size_t size, void *cpu_addr, rt_ubase_t dma_handle, rt_ubase_t flags); rt_inline void region_pool_lock(void) { rt_hw_spin_lock(&dma_pools_lock.lock); } rt_inline void region_pool_unlock(void) { rt_hw_spin_unlock(&dma_pools_lock.lock); } static rt_err_t dma_map_coherent_sync_out_data(struct rt_device *dev, void *data, rt_size_t size, rt_ubase_t *dma_handle, rt_ubase_t flags) { if (dma_handle) { *dma_handle = (rt_ubase_t)rt_kmem_v2p(data); } rt_hw_cpu_dcache_ops(RT_HW_CACHE_FLUSH, data, size); return RT_EOK; } static rt_err_t dma_map_coherent_sync_in_data(struct rt_device *dev, void *out_data, rt_size_t size, rt_ubase_t dma_handle, rt_ubase_t flags) { rt_hw_cpu_dcache_ops(RT_HW_CACHE_INVALIDATE, out_data, size); return RT_EOK; } static const struct rt_dma_map_ops dma_map_coherent_ops = { .sync_out_data = dma_map_coherent_sync_out_data, .sync_in_data = dma_map_coherent_sync_in_data, }; static rt_err_t dma_map_nocoherent_sync_out_data(struct rt_device *dev, void *data, rt_size_t size, rt_ubase_t *dma_handle, rt_ubase_t flags) { if (dma_handle) { *dma_handle = (rt_ubase_t)rt_kmem_v2p(data); } return RT_EOK; } static rt_err_t dma_map_nocoherent_sync_in_data(struct rt_device *dev, void *out_data, rt_size_t size, rt_ubase_t dma_handle, rt_ubase_t flags) { return RT_EOK; } static const struct rt_dma_map_ops dma_map_nocoherent_ops = { .sync_out_data = dma_map_nocoherent_sync_out_data, .sync_in_data = dma_map_nocoherent_sync_in_data, }; #ifdef RT_USING_OFW rt_inline rt_ubase_t ofw_addr_cpu2dma(struct rt_device *dev, rt_ubase_t addr) { return (rt_ubase_t)rt_ofw_translate_cpu2dma(dev->ofw_node, addr); } rt_inline rt_ubase_t ofw_addr_dma2cpu(struct rt_device *dev, rt_ubase_t addr) { return (rt_ubase_t)rt_ofw_translate_dma2cpu(dev->ofw_node, addr); } static void *ofw_dma_map_alloc(struct rt_device *dev, rt_size_t size, rt_ubase_t *dma_handle, rt_ubase_t flags) { void *cpu_addr = dma_alloc(dev, size, dma_handle, flags); if (cpu_addr && dma_handle) { *dma_handle = ofw_addr_cpu2dma(dev, *dma_handle); } return cpu_addr; } static void ofw_dma_map_free(struct rt_device *dev, rt_size_t size, void *cpu_addr, rt_ubase_t dma_handle, rt_ubase_t flags) { dma_handle = ofw_addr_dma2cpu(dev, dma_handle); dma_free(dev, size, cpu_addr, dma_handle, flags); } static rt_err_t ofw_dma_map_sync_out_data(struct rt_device *dev, void *data, rt_size_t size, rt_ubase_t *dma_handle, rt_ubase_t flags) { rt_err_t err; if (flags & RT_DMA_F_NOCACHE) { err = dma_map_nocoherent_sync_out_data(dev, data, size, dma_handle, flags); } else { err = dma_map_coherent_sync_out_data(dev, data, size, dma_handle, flags); } if (!err && dma_handle) { *dma_handle = ofw_addr_cpu2dma(dev, *dma_handle); } return err; } static rt_err_t ofw_dma_map_sync_in_data(struct rt_device *dev, void *out_data, rt_size_t size, rt_ubase_t dma_handle, rt_ubase_t flags) { dma_handle = ofw_addr_dma2cpu(dev, dma_handle); if (flags & RT_DMA_F_NOCACHE) { return dma_map_nocoherent_sync_in_data(dev, out_data, size, dma_handle, flags); } return dma_map_coherent_sync_in_data(dev, out_data, size, dma_handle, flags); } static const struct rt_dma_map_ops ofw_dma_map_ops = { .alloc = ofw_dma_map_alloc, .free = ofw_dma_map_free, .sync_out_data = ofw_dma_map_sync_out_data, .sync_in_data = ofw_dma_map_sync_in_data, }; static const struct rt_dma_map_ops *ofw_device_dma_ops(struct rt_device *dev) { rt_err_t err; int region_nr = 0; const fdt32_t *cell; rt_phandle phandle; rt_region_t region; struct rt_ofw_prop *prop; struct rt_dma_pool *dma_pool; const struct rt_dma_map_ops *ops = RT_NULL; struct rt_ofw_node *mem_np, *np = dev->ofw_node; rt_ofw_foreach_prop_u32(np, "memory-region", prop, cell, phandle) { rt_uint64_t addr, size; if (!(mem_np = rt_ofw_find_node_by_phandle(phandle))) { if (region_nr == 0) { return RT_NULL; } break; } if ((err = rt_ofw_get_address(mem_np, 0, &addr, &size))) { LOG_E("%s: Read '%s' error = %s", rt_ofw_node_full_name(mem_np), "memory-region", rt_strerror(err)); rt_ofw_node_put(mem_np); continue; } region.start = addr; region.end = addr + size; region.name = rt_dm_dev_get_name(dev); rt_ofw_node_put(mem_np); if (!(dma_pool = dma_pool_install(®ion))) { return RT_NULL; } if (rt_ofw_prop_read_bool(mem_np, "no-map")) { dma_pool->flags |= RT_DMA_F_NOMAP; } if (!rt_dma_device_is_coherent(dev)) { dma_pool->flags |= RT_DMA_F_NOCACHE; } dma_pool->dev = dev; ++region_nr; } if (region_nr) { ops = &ofw_dma_map_ops; } return ops; } #endif /* RT_USING_OFW */ static const struct rt_dma_map_ops *device_dma_ops(struct rt_device *dev) { const struct rt_dma_map_ops *ops = dev->dma_ops; if (ops) { return ops; } #ifdef RT_USING_OFW if (dev->ofw_node && (ops = ofw_device_dma_ops(dev))) { return ops; } #endif if (rt_dma_device_is_coherent(dev)) { ops = &dma_map_coherent_ops; } else { ops = &dma_map_nocoherent_ops; } dev->dma_ops = ops; return ops; } static rt_ubase_t dma_pool_alloc(struct rt_dma_pool *pool, rt_size_t size) { rt_size_t bit, next_bit, end_bit, max_bits; size = RT_DIV_ROUND_UP(size, ARCH_PAGE_SIZE); max_bits = pool->bits - size; rt_bitmap_for_each_clear_bit(pool->map, bit, max_bits) { end_bit = bit + size; for (next_bit = bit + 1; next_bit < end_bit; ++next_bit) { if (rt_bitmap_test_bit(pool->map, next_bit)) { bit = next_bit; goto _next; } } if (next_bit == end_bit) { while (next_bit --> bit) { rt_bitmap_set_bit(pool->map, next_bit); } return pool->start + bit * ARCH_PAGE_SIZE; } _next: } return RT_NULL; } static void dma_pool_free(struct rt_dma_pool *pool, rt_ubase_t offset, rt_size_t size) { rt_size_t bit = (offset - pool->start) / ARCH_PAGE_SIZE, end_bit; size = RT_DIV_ROUND_UP(size, ARCH_PAGE_SIZE); end_bit = bit + size; for (; bit < end_bit; ++bit) { rt_bitmap_clear_bit(pool->map, bit); } } static void *dma_alloc(struct rt_device *dev, rt_size_t size, rt_ubase_t *dma_handle, rt_ubase_t flags) { void *dma_buffer = RT_NULL; struct rt_dma_pool *pool; region_pool_lock(); rt_list_for_each_entry(pool, &dma_pool_nodes, list) { if ((flags & RT_DMA_F_DEVICE) && (!(pool->flags & RT_DMA_F_DEVICE) || pool->dev != dev)) { continue; } if ((flags & RT_DMA_F_NOMAP) && !((pool->flags & RT_DMA_F_NOMAP))) { continue; } if ((flags & RT_DMA_F_32BITS) && !((pool->flags & RT_DMA_F_32BITS))) { continue; } if ((flags & RT_DMA_F_LINEAR) && !((pool->flags & RT_DMA_F_LINEAR))) { continue; } *dma_handle = dma_pool_alloc(pool, size); if (*dma_handle && !(flags & RT_DMA_F_NOMAP)) { if (flags & RT_DMA_F_NOCACHE) { dma_buffer = rt_ioremap_nocache((void *)*dma_handle, size); } else { dma_buffer = rt_ioremap_cached((void *)*dma_handle, size); } if (!dma_buffer) { dma_pool_free(pool, *dma_handle, size); continue; } break; } else if (*dma_handle) { dma_buffer = (void *)*dma_handle; break; } } region_pool_unlock(); return dma_buffer; } static void dma_free(struct rt_device *dev, rt_size_t size, void *cpu_addr, rt_ubase_t dma_handle, rt_ubase_t flags) { struct rt_dma_pool *pool; region_pool_lock(); rt_list_for_each_entry(pool, &dma_pool_nodes, list) { if (dma_handle >= pool->region.start && dma_handle <= pool->region.end) { rt_iounmap(cpu_addr); dma_pool_free(pool, dma_handle, size); break; } } region_pool_unlock(); } void *rt_dma_alloc(struct rt_device *dev, rt_size_t size, rt_ubase_t *dma_handle, rt_ubase_t flags) { void *dma_buffer = RT_NULL; rt_ubase_t dma_handle_s = 0; const struct rt_dma_map_ops *ops; if (!dev || !size) { return RT_NULL; } ops = device_dma_ops(dev); if (ops->alloc) { dma_buffer = ops->alloc(dev, size, &dma_handle_s, flags); } else { dma_buffer = dma_alloc(dev, size, &dma_handle_s, flags); } if (!dma_buffer) { return dma_buffer; } if (dma_handle) { *dma_handle = dma_handle_s; } return dma_buffer; } void rt_dma_free(struct rt_device *dev, rt_size_t size, void *cpu_addr, rt_ubase_t dma_handle, rt_ubase_t flags) { const struct rt_dma_map_ops *ops; if (!dev || !size || !cpu_addr) { return; } ops = device_dma_ops(dev); if (ops->free) { ops->free(dev, size, cpu_addr, dma_handle, flags); } else { dma_free(dev, size, cpu_addr, dma_handle, flags); } } rt_err_t rt_dma_sync_out_data(struct rt_device *dev, void *data, rt_size_t size, rt_ubase_t *dma_handle, rt_ubase_t flags) { rt_err_t err; rt_ubase_t dma_handle_s = 0; const struct rt_dma_map_ops *ops; if (!data || !size) { return -RT_EINVAL; } ops = device_dma_ops(dev); err = ops->sync_out_data(dev, data, size, &dma_handle_s, flags); if (dma_handle) { *dma_handle = dma_handle_s; } return err; } rt_err_t rt_dma_sync_in_data(struct rt_device *dev, void *out_data, rt_size_t size, rt_ubase_t dma_handle, rt_ubase_t flags) { rt_err_t err; const struct rt_dma_map_ops *ops; if (!out_data || !size) { return -RT_EINVAL; } ops = device_dma_ops(dev); err = ops->sync_in_data(dev, out_data, size, dma_handle, flags); return err; } static struct rt_dma_pool *dma_pool_install(rt_region_t *region) { rt_err_t err; struct rt_dma_pool *pool; if (!(pool = rt_calloc(1, sizeof(*pool)))) { LOG_E("Install pool[%p, %p] error = %s", region->start, region->end, rt_strerror(-RT_ENOMEM)); return RT_NULL; } rt_memcpy(&pool->region, region, sizeof(*region)); pool->flags |= RT_DMA_F_LINEAR; if (region->end < 4UL * SIZE_GB) { pool->flags |= RT_DMA_F_32BITS; } pool->start = RT_ALIGN(pool->region.start, ARCH_PAGE_SIZE); pool->bits = (pool->region.end - pool->start) / ARCH_PAGE_SIZE; if (!pool->bits) { err = -RT_EINVAL; goto _fail; } pool->map = rt_calloc(RT_BITMAP_LEN(pool->bits), sizeof(*pool->map)); if (!pool->map) { err = -RT_ENOMEM; goto _fail; } rt_list_init(&pool->list); region_pool_lock(); rt_list_insert_before(&dma_pool_nodes, &pool->list); region_pool_unlock(); return pool; _fail: rt_free(pool); LOG_E("Install pool[%p, %p] error = %s", region->start, region->end, rt_strerror(err)); return RT_NULL; } struct rt_dma_pool *rt_dma_pool_install(rt_region_t *region) { struct rt_dma_pool *pool; if (!region) { return RT_NULL; } if ((pool = dma_pool_install(region))) { region = &pool->region; LOG_I("%s: Reserved %u.%u MiB at %p", region->name, (region->end - region->start) / SIZE_MB, (region->end - region->start) / SIZE_KB & (SIZE_KB - 1), region->start); } return pool; } rt_err_t rt_dma_pool_extract(rt_region_t *region_list, rt_size_t list_len, rt_size_t cma_size, rt_size_t coherent_pool_size) { struct rt_dma_pool *pool; rt_region_t *region = region_list, *region_high = RT_NULL, cma, coherent_pool; if (!region_list || !list_len || cma_size < coherent_pool_size) { return -RT_EINVAL; } for (rt_size_t i = 0; i < list_len; ++i, ++region) { if (!region->name) { continue; } /* Always use low address in 4G */ if (region->end - region->start >= cma_size) { if ((rt_ssize_t)((4UL * SIZE_GB) - region->start) < cma_size) { region_high = region; continue; } goto _found; } } if (region_high) { region = region_high; LOG_W("No available DMA zone in 4G"); goto _found; } return -RT_EEMPTY; _found: if (region->end - region->start != cma_size) { cma.start = region->start; cma.end = cma.start + cma_size; /* Update input region */ region->start += cma_size; } else { rt_memcpy(&cma, region, sizeof(cma)); } coherent_pool.name = "coherent-pool"; coherent_pool.start = cma.start; coherent_pool.end = coherent_pool.start + coherent_pool_size; cma.name = "cma"; cma.start += coherent_pool_size; if (!(pool = rt_dma_pool_install(&coherent_pool))) { return -RT_ENOMEM; } /* Use: CMA > coherent-pool */ if (!(pool = rt_dma_pool_install(&cma))) { return -RT_ENOMEM; } return RT_EOK; } #if defined(RT_USING_CONSOLE) && defined(RT_USING_MSH) static int list_dma_pool(int argc, char**argv) { int count = 0; rt_region_t *region; struct rt_dma_pool *pool; rt_kprintf("%-*.s Region\n", RT_NAME_MAX, "Name"); region_pool_lock(); rt_list_for_each_entry(pool, &dma_pool_nodes, list) { region = &pool->region; rt_kprintf("%-*.s [%p, %p]\n", RT_NAME_MAX, region->name, region->start, region->end); ++count; } rt_kprintf("%d DMA memory found\n", count); region_pool_unlock(); return 0; } MSH_CMD_EXPORT(list_dma_pool, dump all dma memory pool); #endif /* RT_USING_CONSOLE && RT_USING_MSH */