| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497 |
- /*
- * Copyright (C) 2022 Tencent Corporation. All rights reserved.
- * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
- */
- #include "ems_gc.h"
- #include "ems_gc_internal.h"
- #define GB (1 << 30UL)
- #define MARK_NODE_OBJ_CNT 256
- #if WASM_ENABLE_GC != 0
- /* mark node is used for gc marker*/
- typedef struct mark_node_struct {
- /* number of to-expand objects can be saved in this node */
- gc_size_t cnt;
- /* the first unused index */
- uint32 idx;
- /* next node on the node list */
- struct mark_node_struct *next;
- /* the actual to-expand objects list */
- gc_object_t set[MARK_NODE_OBJ_CNT];
- } mark_node_t;
- /**
- * Alloc a mark node from the native heap
- *
- * @return a valid mark node if success, NULL otherwise
- */
- static mark_node_t *
- alloc_mark_node(void)
- {
- mark_node_t *ret = (mark_node_t *)BH_MALLOC(sizeof(mark_node_t));
- if (!ret) {
- LOG_ERROR("alloc a new mark node failed");
- return NULL;
- }
- ret->cnt = sizeof(ret->set) / sizeof(ret->set[0]);
- ret->idx = 0;
- ret->next = NULL;
- return ret;
- }
- /* Free a mark node to the native heap
- *
- * @param node the mark node to free, should not be NULL
- */
- static void
- free_mark_node(mark_node_t *node)
- {
- bh_assert(node);
- BH_FREE((gc_object_t)node);
- }
- /**
- * Sweep phase of mark_sweep algorithm
- * @param heap the heap to sweep, should be a valid instance heap
- * which has already been marked
- */
- static void
- sweep_instance_heap(gc_heap_t *heap)
- {
- hmu_t *cur = NULL, *end = NULL, *last = NULL;
- hmu_type_t ut;
- gc_size_t size;
- int i, lsize;
- gc_size_t tot_free = 0;
- bh_assert(gci_is_heap_valid(heap));
- cur = (hmu_t *)heap->base_addr;
- last = NULL;
- end = (hmu_t *)((char *)heap->base_addr + heap->current_size);
- /* reset KFC */
- lsize =
- (int)(sizeof(heap->kfc_normal_list) / sizeof(heap->kfc_normal_list[0]));
- for (i = 0; i < lsize; i++) {
- heap->kfc_normal_list[i].next = NULL;
- }
- heap->kfc_tree_root->right = NULL;
- heap->root_set = NULL;
- while (cur < end) {
- ut = hmu_get_ut(cur);
- size = hmu_get_size(cur);
- bh_assert(size > 0);
- if (ut == HMU_FC || ut == HMU_FM
- || (ut == HMU_VO && hmu_is_vo_freed(cur))
- || (ut == HMU_WO && !hmu_is_wo_marked(cur))) {
- /* merge previous free areas with current one */
- if (!last)
- last = cur;
- if (ut == HMU_WO) {
- /* Invoke registered finalizer */
- gc_object_t cur_obj = hmu_to_obj(cur);
- if (gct_vm_get_extra_info_flag(cur_obj)) {
- extra_info_node_t *node = gc_search_extra_info_node(
- (gc_handle_t)heap, cur_obj, NULL);
- bh_assert(node);
- node->finalizer(node->obj, node->data);
- gc_unset_finalizer((gc_handle_t)heap, cur_obj);
- }
- }
- }
- else {
- /* current block is still live */
- if (last) {
- tot_free += (gc_size_t)((char *)cur - (char *)last);
- gci_add_fc(heap, last, (gc_size_t)((char *)cur - (char *)last));
- hmu_mark_pinuse(last);
- last = NULL;
- }
- if (ut == HMU_WO) {
- /* unmark it */
- hmu_unmark_wo(cur);
- }
- }
- cur = (hmu_t *)((char *)cur + size);
- }
- bh_assert(cur == end);
- if (last) {
- tot_free += (gc_size_t)((char *)cur - (char *)last);
- gci_add_fc(heap, last, (gc_size_t)((char *)cur - (char *)last));
- hmu_mark_pinuse(last);
- }
- heap->total_free_size = tot_free;
- #if GC_STAT_DATA != 0
- heap->total_gc_count++;
- if ((heap->current_size - tot_free) > heap->highmark_size)
- heap->highmark_size = heap->current_size - tot_free;
- #endif
- gc_update_threshold(heap);
- }
- /**
- * Add a to-expand node to the to-expand list
- *
- * @param heap should be a valid instance heap
- * @param obj should be a valid wo inside @heap
- *
- * @return GC_ERROR if there is no more resource for marking,
- * GC_SUCCESS if success
- */
- static int
- add_wo_to_expand(gc_heap_t *heap, gc_object_t obj)
- {
- mark_node_t *mark_node = NULL, *new_node = NULL;
- hmu_t *hmu = NULL;
- bh_assert(obj);
- hmu = obj_to_hmu(obj);
- bh_assert(gci_is_heap_valid(heap));
- bh_assert((gc_uint8 *)hmu >= heap->base_addr
- && (gc_uint8 *)hmu < heap->base_addr + heap->current_size);
- bh_assert(hmu_get_ut(hmu) == HMU_WO);
- if (hmu_is_wo_marked(hmu))
- return GC_SUCCESS; /* already marked*/
- mark_node = (mark_node_t *)heap->root_set;
- if (!mark_node || mark_node->idx == mark_node->cnt) {
- new_node = alloc_mark_node();
- if (!new_node) {
- LOG_ERROR("can not add obj to mark node because of mark node "
- "allocation failed");
- return GC_ERROR;
- }
- new_node->next = mark_node;
- heap->root_set = new_node;
- mark_node = new_node;
- }
- mark_node->set[mark_node->idx++] = obj;
- hmu_mark_wo(hmu);
- return GC_SUCCESS;
- }
- /* Check ems_gc.h for description*/
- int
- gc_add_root(void *heap_p, gc_object_t obj)
- {
- gc_heap_t *heap = (gc_heap_t *)heap_p;
- hmu_t *hmu = NULL;
- if (!obj) {
- LOG_ERROR("gc_add_root with NULL obj");
- return GC_ERROR;
- }
- hmu = obj_to_hmu(obj);
- if (!gci_is_heap_valid(heap)) {
- LOG_ERROR("vm_get_gc_handle_for_current_instance returns invalid heap");
- return GC_ERROR;
- }
- if (!((gc_uint8 *)hmu >= heap->base_addr
- && (gc_uint8 *)hmu < heap->base_addr + heap->current_size)) {
- LOG_ERROR("Obj is not a object in current instance heap");
- return GC_ERROR;
- }
- if (hmu_get_ut(hmu) != HMU_WO) {
- LOG_ERROR("Given object is not wo");
- return GC_ERROR;
- }
- if (add_wo_to_expand(heap, obj) != GC_SUCCESS) {
- heap->is_fast_marking_failed = 1;
- return GC_ERROR;
- }
- return GC_SUCCESS;
- }
- /**
- * Unmark all marked objects to do rollback
- *
- * @param heap the heap to do rollback, should be a valid instance heap
- */
- static void
- rollback_mark(gc_heap_t *heap)
- {
- mark_node_t *mark_node = NULL, *next_mark_node = NULL;
- hmu_t *cur = NULL, *end = NULL;
- hmu_type_t ut;
- gc_size_t size;
- bh_assert(gci_is_heap_valid(heap));
- /* roll back*/
- mark_node = (mark_node_t *)heap->root_set;
- while (mark_node) {
- next_mark_node = mark_node->next;
- free_mark_node(mark_node);
- mark_node = next_mark_node;
- }
- heap->root_set = NULL;
- /* then traverse the heap to unmark all marked wos*/
- cur = (hmu_t *)heap->base_addr;
- end = (hmu_t *)((char *)heap->base_addr + heap->current_size);
- while (cur < end) {
- ut = hmu_get_ut(cur);
- size = hmu_get_size(cur);
- if (ut == HMU_WO && hmu_is_wo_marked(cur)) {
- hmu_unmark_wo(cur);
- }
- cur = (hmu_t *)((char *)cur + size);
- }
- bh_assert(cur == end);
- }
- /**
- * Reclaim GC instance heap
- *
- * @param heap the heap to reclaim, should be a valid instance heap
- *
- * @return GC_SUCCESS if success, GC_ERROR otherwise
- */
- static int
- reclaim_instance_heap(gc_heap_t *heap)
- {
- mark_node_t *mark_node = NULL;
- int idx = 0, j = 0;
- bool ret, is_compact_mode = false;
- gc_object_t obj = NULL, ref = NULL;
- hmu_t *hmu = NULL;
- gc_uint32 ref_num = 0, ref_start_offset = 0, size = 0, offset = 0;
- gc_uint16 *ref_list = NULL;
- bh_assert(gci_is_heap_valid(heap));
- heap->root_set = NULL;
- #if WASM_ENABLE_THREAD_MGR == 0
- if (!heap->exec_env)
- return GC_SUCCESS;
- ret = gct_vm_begin_rootset_enumeration(heap->exec_env, heap);
- #else
- if (!heap->cluster)
- return GC_SUCCESS;
- ret = gct_vm_begin_rootset_enumeration(heap->cluster, heap);
- #endif
- if (!ret)
- return GC_ERROR;
- #if BH_ENABLE_GC_VERIFY != 0
- /* no matter whether the enumeration is successful or not, the data
- collected should be checked at first */
- mark_node = (mark_node_t *)heap->root_set;
- while (mark_node) {
- /* all nodes except first should be full filled */
- bh_assert(mark_node == (mark_node_t *)heap->root_set
- || mark_node->idx == mark_node->cnt);
- /* all nodes should be non-empty */
- bh_assert(mark_node->idx > 0);
- for (idx = 0; idx < (int)mark_node->idx; idx++) {
- obj = mark_node->set[idx];
- hmu = obj_to_hmu(obj);
- bh_assert(hmu_is_wo_marked(hmu));
- bh_assert((gc_uint8 *)hmu >= heap->base_addr
- && (gc_uint8 *)hmu
- < heap->base_addr + heap->current_size);
- }
- mark_node = mark_node->next;
- }
- #endif
- /* TODO: when fast marking failed, we can still do slow
- marking, currently just simply roll it back. */
- if (heap->is_fast_marking_failed) {
- LOG_ERROR("enumerate rootset failed");
- LOG_ERROR("all marked wos will be unmarked to keep heap consistency");
- rollback_mark(heap);
- heap->is_fast_marking_failed = 0;
- return GC_ERROR;
- }
- /* the algorithm we use to mark all objects */
- /* 1. mark rootset and organize them into a mark_node list (last marked
- * roots at list header, i.e. stack top) */
- /* 2. in every iteration, we use the top node to expand*/
- /* 3. execute step 2 till no expanding */
- /* this is a BFS & DFS mixed algorithm, but more like DFS */
- mark_node = (mark_node_t *)heap->root_set;
- while (mark_node) {
- heap->root_set = mark_node->next;
- /* note that mark_node->idx may change in each loop */
- for (idx = 0; idx < (int)mark_node->idx; idx++) {
- obj = mark_node->set[idx];
- hmu = obj_to_hmu(obj);
- size = hmu_get_size(hmu);
- if (!gct_vm_get_wasm_object_ref_list(obj, &is_compact_mode,
- &ref_num, &ref_list,
- &ref_start_offset)) {
- LOG_ERROR("mark process failed because failed "
- "vm_get_wasm_object_ref_list");
- break;
- }
- if (ref_num >= 2U * GB) {
- LOG_ERROR("Invalid ref_num returned");
- break;
- }
- if (is_compact_mode) {
- for (j = 0; j < (int)ref_num; j++) {
- offset = ref_start_offset + j * sizeof(void *);
- bh_assert(offset + sizeof(void *) < size);
- ref = *(gc_object_t *)(((gc_uint8 *)obj) + offset);
- if (ref == NULL_REF || ((uintptr_t)ref & 1))
- continue; /* null object or i31 object */
- if (add_wo_to_expand(heap, ref) == GC_ERROR) {
- LOG_ERROR("add_wo_to_expand failed");
- break;
- }
- }
- if (j < (int)ref_num)
- break;
- }
- else {
- for (j = 0; j < (int)ref_num; j++) {
- offset = ref_list[j];
- bh_assert(offset + sizeof(void *) < size);
- ref = *(gc_object_t *)(((gc_uint8 *)obj) + offset);
- if (ref == NULL_REF || ((uintptr_t)ref & 1))
- continue; /* null object or i31 object */
- if (add_wo_to_expand(heap, ref) == GC_ERROR) {
- LOG_ERROR("mark process failed");
- break;
- }
- }
- if (j < (int)ref_num)
- break;
- }
- }
- if (idx < (int)mark_node->idx)
- break; /* not yet done */
- /* obj's in mark_node are all expanded */
- free_mark_node(mark_node);
- mark_node = heap->root_set;
- }
- if (mark_node) {
- LOG_ERROR("mark process is not successfully finished");
- free_mark_node(mark_node);
- /* roll back is required */
- rollback_mark(heap);
- return GC_ERROR;
- }
- /* now sweep */
- sweep_instance_heap(heap);
- (void)size;
- return GC_SUCCESS;
- }
- /**
- * Do GC on given heap
- *
- * @param the heap to do GC, should be a valid heap
- *
- * @return GC_SUCCESS if success, GC_ERROR otherwise
- */
- int
- gci_gc_heap(void *h)
- {
- int ret = GC_ERROR;
- gc_heap_t *heap = (gc_heap_t *)h;
- bh_assert(gci_is_heap_valid(heap));
- LOG_VERBOSE("#reclaim instance heap %p", heap);
- /* TODO: get exec_env of current thread when GC multi-threading
- is enabled, and pass it to runtime */
- gct_vm_gc_prepare(NULL);
- gct_vm_mutex_lock(&heap->lock);
- heap->is_doing_reclaim = 1;
- ret = reclaim_instance_heap(heap);
- heap->is_doing_reclaim = 0;
- gct_vm_mutex_unlock(&heap->lock);
- /* TODO: get exec_env of current thread when GC multi-threading
- is enabled, and pass it to runtime */
- gct_vm_gc_finished(NULL);
- LOG_VERBOSE("#reclaim instance heap %p done", heap);
- #if BH_ENABLE_GC_VERIFY != 0
- gci_verify_heap(heap);
- #endif
- #if GC_STAT_SHOW != 0
- gc_show_stat(heap);
- gc_show_fragment(heap);
- #endif
- return ret;
- }
- int
- gc_is_dead_object(void *obj)
- {
- return !hmu_is_wo_marked(obj_to_hmu(obj));
- }
- #else
- int
- gci_gc_heap(void *h)
- {
- (void)h;
- return GC_ERROR;
- }
- #endif /* end of WASM_ENABLE_GC != 0 */
|