| /* |
| * SPDX-License-Identifier: BSD-2-Clause |
| * |
| * Copyright (c) 2023 Ventana Micro Systems Inc. |
| * |
| * Authors: |
| * Anup Patel<apatel@ventanamicro.com> |
| */ |
| |
| #include <sbi/riscv_locks.h> |
| #include <sbi/sbi_error.h> |
| #include <sbi/sbi_heap.h> |
| #include <sbi/sbi_list.h> |
| #include <sbi/sbi_scratch.h> |
| #include <sbi/sbi_string.h> |
| |
| /* Alignment of heap base address and size */ |
| #define HEAP_BASE_ALIGN 1024 |
| /* Minimum size and alignment of heap allocations */ |
| #define HEAP_ALLOC_ALIGN 64 |
| #define HEAP_HOUSEKEEPING_FACTOR 16 |
| |
| struct heap_node { |
| struct sbi_dlist head; |
| unsigned long addr; |
| unsigned long size; |
| }; |
| |
| struct heap_control { |
| spinlock_t lock; |
| unsigned long base; |
| unsigned long size; |
| unsigned long hkbase; |
| unsigned long hksize; |
| struct sbi_dlist free_node_list; |
| struct sbi_dlist free_space_list; |
| struct sbi_dlist used_space_list; |
| }; |
| |
| static struct heap_control hpctrl; |
| |
| void *sbi_malloc(size_t size) |
| { |
| void *ret = NULL; |
| struct heap_node *n, *np; |
| |
| if (!size) |
| return NULL; |
| |
| size += HEAP_ALLOC_ALIGN - 1; |
| size &= ~((unsigned long)HEAP_ALLOC_ALIGN - 1); |
| |
| spin_lock(&hpctrl.lock); |
| |
| np = NULL; |
| sbi_list_for_each_entry(n, &hpctrl.free_space_list, head) { |
| if (size <= n->size) { |
| np = n; |
| break; |
| } |
| } |
| if (np) { |
| if ((size < np->size) && |
| !sbi_list_empty(&hpctrl.free_node_list)) { |
| n = sbi_list_first_entry(&hpctrl.free_node_list, |
| struct heap_node, head); |
| sbi_list_del(&n->head); |
| n->addr = np->addr + np->size - size; |
| n->size = size; |
| np->size -= size; |
| sbi_list_add_tail(&n->head, &hpctrl.used_space_list); |
| ret = (void *)n->addr; |
| } else if (size == np->size) { |
| sbi_list_del(&np->head); |
| sbi_list_add_tail(&np->head, &hpctrl.used_space_list); |
| ret = (void *)np->addr; |
| } |
| } |
| |
| spin_unlock(&hpctrl.lock); |
| |
| return ret; |
| } |
| |
| void *sbi_zalloc(size_t size) |
| { |
| void *ret = sbi_malloc(size); |
| |
| if (ret) |
| sbi_memset(ret, 0, size); |
| return ret; |
| } |
| |
| void sbi_free(void *ptr) |
| { |
| struct heap_node *n, *np; |
| |
| if (!ptr) |
| return; |
| |
| spin_lock(&hpctrl.lock); |
| |
| np = NULL; |
| sbi_list_for_each_entry(n, &hpctrl.used_space_list, head) { |
| if ((n->addr <= (unsigned long)ptr) && |
| ((unsigned long)ptr < (n->addr + n->size))) { |
| np = n; |
| break; |
| } |
| } |
| if (!np) { |
| spin_unlock(&hpctrl.lock); |
| return; |
| } |
| |
| sbi_list_del(&np->head); |
| |
| sbi_list_for_each_entry(n, &hpctrl.free_space_list, head) { |
| if ((np->addr + np->size) == n->addr) { |
| n->addr = np->addr; |
| n->size += np->size; |
| sbi_list_add_tail(&np->head, &hpctrl.free_node_list); |
| np = NULL; |
| break; |
| } else if (np->addr == (n->addr + n->size)) { |
| n->size += np->size; |
| sbi_list_add_tail(&np->head, &hpctrl.free_node_list); |
| np = NULL; |
| break; |
| } else if ((n->addr + n->size) < np->addr) { |
| sbi_list_add(&np->head, &n->head); |
| np = NULL; |
| break; |
| } |
| } |
| if (np) |
| sbi_list_add_tail(&np->head, &hpctrl.free_space_list); |
| |
| spin_unlock(&hpctrl.lock); |
| } |
| |
| unsigned long sbi_heap_free_space(void) |
| { |
| struct heap_node *n; |
| unsigned long ret = 0; |
| |
| spin_lock(&hpctrl.lock); |
| sbi_list_for_each_entry(n, &hpctrl.free_space_list, head) |
| ret += n->size; |
| spin_unlock(&hpctrl.lock); |
| |
| return ret; |
| } |
| |
| unsigned long sbi_heap_used_space(void) |
| { |
| return hpctrl.size - hpctrl.hksize - sbi_heap_free_space(); |
| } |
| |
| unsigned long sbi_heap_reserved_space(void) |
| { |
| return hpctrl.hksize; |
| } |
| |
| int sbi_heap_init(struct sbi_scratch *scratch) |
| { |
| unsigned long i; |
| struct heap_node *n; |
| |
| /* Sanity checks on heap offset and size */ |
| if (!scratch->fw_heap_size || |
| (scratch->fw_heap_size & (HEAP_BASE_ALIGN - 1)) || |
| (scratch->fw_heap_offset < scratch->fw_rw_offset) || |
| (scratch->fw_size < (scratch->fw_heap_offset + scratch->fw_heap_size)) || |
| (scratch->fw_heap_offset & (HEAP_BASE_ALIGN - 1))) |
| return SBI_EINVAL; |
| |
| /* Initialize heap control */ |
| SPIN_LOCK_INIT(hpctrl.lock); |
| hpctrl.base = scratch->fw_start + scratch->fw_heap_offset; |
| hpctrl.size = scratch->fw_heap_size; |
| hpctrl.hkbase = hpctrl.base; |
| hpctrl.hksize = hpctrl.size / HEAP_HOUSEKEEPING_FACTOR; |
| hpctrl.hksize &= ~((unsigned long)HEAP_BASE_ALIGN - 1); |
| SBI_INIT_LIST_HEAD(&hpctrl.free_node_list); |
| SBI_INIT_LIST_HEAD(&hpctrl.free_space_list); |
| SBI_INIT_LIST_HEAD(&hpctrl.used_space_list); |
| |
| /* Prepare free node list */ |
| for (i = 0; i < (hpctrl.hksize / sizeof(*n)); i++) { |
| n = (struct heap_node *)(hpctrl.hkbase + (sizeof(*n) * i)); |
| SBI_INIT_LIST_HEAD(&n->head); |
| n->addr = n->size = 0; |
| sbi_list_add_tail(&n->head, &hpctrl.free_node_list); |
| } |
| |
| /* Prepare free space list */ |
| n = sbi_list_first_entry(&hpctrl.free_node_list, |
| struct heap_node, head); |
| sbi_list_del(&n->head); |
| n->addr = hpctrl.hkbase + hpctrl.hksize; |
| n->size = hpctrl.size - hpctrl.hksize; |
| sbi_list_add_tail(&n->head, &hpctrl.free_space_list); |
| |
| return 0; |
| } |