mirror of
https://gitee.com/bianbu-linux/opensbi
synced 2025-04-24 23:07:08 -04:00
206 lines
4.6 KiB
C
206 lines
4.6 KiB
C
/*
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
|
*
|
|
* Copyright (c) 2023 Ventana Micro Systems Inc.
|
|
*
|
|
* Authors:
|
|
* Anup Patel<apatel@ventanamicro.com>
|
|
*/
|
|
|
|
#include <sbi/riscv_locks.h>
|
|
#include <sbi/sbi_error.h>
|
|
#include <sbi/sbi_heap.h>
|
|
#include <sbi/sbi_list.h>
|
|
#include <sbi/sbi_scratch.h>
|
|
#include <sbi/sbi_string.h>
|
|
|
|
/* Alignment of heap base address and size */
|
|
#define HEAP_BASE_ALIGN 1024
|
|
/* Minimum size and alignment of heap allocations */
|
|
#define HEAP_ALLOC_ALIGN 64
|
|
#define HEAP_HOUSEKEEPING_FACTOR 16
|
|
|
|
struct heap_node {
|
|
struct sbi_dlist head;
|
|
unsigned long addr;
|
|
unsigned long size;
|
|
};
|
|
|
|
struct heap_control {
|
|
spinlock_t lock;
|
|
unsigned long base;
|
|
unsigned long size;
|
|
unsigned long hkbase;
|
|
unsigned long hksize;
|
|
struct sbi_dlist free_node_list;
|
|
struct sbi_dlist free_space_list;
|
|
struct sbi_dlist used_space_list;
|
|
};
|
|
|
|
static struct heap_control hpctrl;
|
|
|
|
void *sbi_malloc(size_t size)
|
|
{
|
|
void *ret = NULL;
|
|
struct heap_node *n, *np;
|
|
|
|
if (!size)
|
|
return NULL;
|
|
|
|
size += HEAP_ALLOC_ALIGN - 1;
|
|
size &= ~((unsigned long)HEAP_ALLOC_ALIGN - 1);
|
|
|
|
spin_lock(&hpctrl.lock);
|
|
|
|
np = NULL;
|
|
sbi_list_for_each_entry(n, &hpctrl.free_space_list, head) {
|
|
if (size <= n->size) {
|
|
np = n;
|
|
break;
|
|
}
|
|
}
|
|
if (np) {
|
|
if ((size < np->size) &&
|
|
!sbi_list_empty(&hpctrl.free_node_list)) {
|
|
n = sbi_list_first_entry(&hpctrl.free_node_list,
|
|
struct heap_node, head);
|
|
sbi_list_del(&n->head);
|
|
n->addr = np->addr + np->size - size;
|
|
n->size = size;
|
|
np->size -= size;
|
|
sbi_list_add_tail(&n->head, &hpctrl.used_space_list);
|
|
ret = (void *)n->addr;
|
|
} else if (size == np->size) {
|
|
sbi_list_del(&np->head);
|
|
sbi_list_add_tail(&np->head, &hpctrl.used_space_list);
|
|
ret = (void *)np->addr;
|
|
}
|
|
}
|
|
|
|
spin_unlock(&hpctrl.lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
void *sbi_zalloc(size_t size)
|
|
{
|
|
void *ret = sbi_malloc(size);
|
|
|
|
if (ret)
|
|
sbi_memset(ret, 0, size);
|
|
return ret;
|
|
}
|
|
|
|
void sbi_free(void *ptr)
|
|
{
|
|
struct heap_node *n, *np;
|
|
|
|
if (!ptr)
|
|
return;
|
|
|
|
spin_lock(&hpctrl.lock);
|
|
|
|
np = NULL;
|
|
sbi_list_for_each_entry(n, &hpctrl.used_space_list, head) {
|
|
if ((n->addr <= (unsigned long)ptr) &&
|
|
((unsigned long)ptr < (n->addr + n->size))) {
|
|
np = n;
|
|
break;
|
|
}
|
|
}
|
|
if (!np) {
|
|
spin_unlock(&hpctrl.lock);
|
|
return;
|
|
}
|
|
|
|
sbi_list_del(&np->head);
|
|
|
|
sbi_list_for_each_entry(n, &hpctrl.free_space_list, head) {
|
|
if ((np->addr + np->size) == n->addr) {
|
|
n->addr = np->addr;
|
|
n->size += np->size;
|
|
sbi_list_add_tail(&np->head, &hpctrl.free_node_list);
|
|
np = NULL;
|
|
break;
|
|
} else if (np->addr == (n->addr + n->size)) {
|
|
n->size += np->size;
|
|
sbi_list_add_tail(&np->head, &hpctrl.free_node_list);
|
|
np = NULL;
|
|
break;
|
|
} else if ((n->addr + n->size) < np->addr) {
|
|
sbi_list_add(&np->head, &n->head);
|
|
np = NULL;
|
|
break;
|
|
}
|
|
}
|
|
if (np)
|
|
sbi_list_add_tail(&np->head, &hpctrl.free_space_list);
|
|
|
|
spin_unlock(&hpctrl.lock);
|
|
}
|
|
|
|
unsigned long sbi_heap_free_space(void)
|
|
{
|
|
struct heap_node *n;
|
|
unsigned long ret = 0;
|
|
|
|
spin_lock(&hpctrl.lock);
|
|
sbi_list_for_each_entry(n, &hpctrl.free_space_list, head)
|
|
ret += n->size;
|
|
spin_unlock(&hpctrl.lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
unsigned long sbi_heap_used_space(void)
|
|
{
|
|
return hpctrl.size - hpctrl.hksize - sbi_heap_free_space();
|
|
}
|
|
|
|
unsigned long sbi_heap_reserved_space(void)
|
|
{
|
|
return hpctrl.hksize;
|
|
}
|
|
|
|
int sbi_heap_init(struct sbi_scratch *scratch)
|
|
{
|
|
unsigned long i;
|
|
struct heap_node *n;
|
|
|
|
/* Sanity checks on heap offset and size */
|
|
if (!scratch->fw_heap_size ||
|
|
(scratch->fw_heap_size & (HEAP_BASE_ALIGN - 1)) ||
|
|
(scratch->fw_heap_offset < scratch->fw_rw_offset) ||
|
|
(scratch->fw_size < (scratch->fw_heap_offset + scratch->fw_heap_size)) ||
|
|
(scratch->fw_heap_offset & (HEAP_BASE_ALIGN - 1)))
|
|
return SBI_EINVAL;
|
|
|
|
/* Initialize heap control */
|
|
SPIN_LOCK_INIT(hpctrl.lock);
|
|
hpctrl.base = scratch->fw_start + scratch->fw_heap_offset;
|
|
hpctrl.size = scratch->fw_heap_size;
|
|
hpctrl.hkbase = hpctrl.base;
|
|
hpctrl.hksize = hpctrl.size / HEAP_HOUSEKEEPING_FACTOR;
|
|
hpctrl.hksize &= ~((unsigned long)HEAP_BASE_ALIGN - 1);
|
|
SBI_INIT_LIST_HEAD(&hpctrl.free_node_list);
|
|
SBI_INIT_LIST_HEAD(&hpctrl.free_space_list);
|
|
SBI_INIT_LIST_HEAD(&hpctrl.used_space_list);
|
|
|
|
/* Prepare free node list */
|
|
for (i = 0; i < (hpctrl.hksize / sizeof(*n)); i++) {
|
|
n = (struct heap_node *)(hpctrl.hkbase + (sizeof(*n) * i));
|
|
SBI_INIT_LIST_HEAD(&n->head);
|
|
n->addr = n->size = 0;
|
|
sbi_list_add_tail(&n->head, &hpctrl.free_node_list);
|
|
}
|
|
|
|
/* Prepare free space list */
|
|
n = sbi_list_first_entry(&hpctrl.free_node_list,
|
|
struct heap_node, head);
|
|
sbi_list_del(&n->head);
|
|
n->addr = hpctrl.hkbase + hpctrl.hksize;
|
|
n->size = hpctrl.size - hpctrl.hksize;
|
|
sbi_list_add_tail(&n->head, &hpctrl.free_space_list);
|
|
|
|
return 0;
|
|
}
|