use super::VMArrayRef;
use super::free_list::FreeList;
use crate::hash_map::HashMap;
use crate::hash_set::HashSet;
use crate::runtime::vm::{
ExternRefHostDataId, ExternRefHostDataTable, GarbageCollection, GcHeap, GcHeapObject,
GcProgress, GcRootsIter, GcRuntime, TypedGcRef, VMExternRef, VMGcHeader, VMGcRef,
};
use crate::vm::VMMemoryDefinition;
use crate::{Engine, EngineWeak, prelude::*};
use core::sync::atomic::AtomicUsize;
use core::{
alloc::Layout,
any::Any,
mem,
ops::{Deref, DerefMut},
ptr::NonNull,
};
use wasmtime_environ::drc::{ARRAY_LENGTH_OFFSET, DrcTypeLayouts};
use wasmtime_environ::{
GcArrayLayout, GcLayout, GcStructLayout, GcTypeLayouts, POISON, VMGcKind, VMSharedTypeIndex,
gc_assert,
};
#[expect(clippy::cast_possible_truncation, reason = "known to not overflow")]
const GC_REF_ARRAY_ELEMS_OFFSET: u32 = ARRAY_LENGTH_OFFSET + (mem::size_of::<u32>() as u32);
#[derive(Default)]
pub struct DrcCollector {
layouts: DrcTypeLayouts,
}
unsafe impl GcRuntime for DrcCollector {
fn layouts(&self) -> &dyn GcTypeLayouts {
&self.layouts
}
fn new_gc_heap(&self, engine: &Engine) -> Result<Box<dyn GcHeap>> {
let heap = DrcHeap::new(engine)?;
Ok(Box::new(heap) as _)
}
}
enum TraceInfo {
Array {
gc_ref_elems: bool,
},
Struct {
gc_ref_offsets: Box<[u32]>,
},
}
struct DrcHeap {
engine: EngineWeak,
trace_infos: HashMap<VMSharedTypeIndex, TraceInfo>,
no_gc_count: u64,
over_approximated_stack_roots: Box<Option<VMGcRef>>,
memory: Option<crate::vm::Memory>,
vmmemory: Option<VMMemoryDefinition>,
free_list: Option<FreeList>,
dec_ref_stack: Option<Vec<VMGcRef>>,
}
impl DrcHeap {
fn new(engine: &Engine) -> Result<Self> {
log::trace!("allocating new DRC heap");
Ok(Self {
engine: engine.weak(),
trace_infos: HashMap::with_capacity(1),
no_gc_count: 0,
over_approximated_stack_roots: Box::new(None),
memory: None,
vmmemory: None,
free_list: None,
dec_ref_stack: Some(Vec::with_capacity(1)),
})
}
fn engine(&self) -> Engine {
self.engine.upgrade().unwrap()
}
fn dealloc(&mut self, gc_ref: VMGcRef) {
let drc_ref = drc_ref(&gc_ref);
let size = self.index(drc_ref).object_size();
let layout = FreeList::layout(size);
let index = gc_ref.as_heap_index().unwrap();
if cfg!(gc_zeal) {
let index = usize::try_from(index.get()).unwrap();
self.heap_slice_mut()[index..][..layout.size()].fill(POISON);
}
self.free_list.as_mut().unwrap().dealloc(index, layout);
}
fn inc_ref(&mut self, gc_ref: &VMGcRef) {
if gc_ref.is_i31() {
return;
}
let drc_ref = drc_ref(gc_ref);
let header = self.index_mut(&drc_ref);
debug_assert_ne!(
header.ref_count, 0,
"{:#p} is supposedly live; should have nonzero ref count",
*gc_ref
);
header.ref_count += 1;
log::trace!("increment {:#p} ref count -> {}", *gc_ref, header.ref_count);
}
fn dec_ref(&mut self, gc_ref: &VMGcRef) -> bool {
if gc_ref.is_i31() {
return false;
}
let drc_ref = drc_ref(gc_ref);
let header = self.index_mut(drc_ref);
debug_assert_ne!(
header.ref_count, 0,
"{:#p} is supposedly live; should have nonzero ref count",
*gc_ref
);
header.ref_count -= 1;
log::trace!("decrement {:#p} ref count -> {}", *gc_ref, header.ref_count);
header.ref_count == 0
}
fn dec_ref_and_maybe_dealloc(
&mut self,
host_data_table: &mut ExternRefHostDataTable,
gc_ref: &VMGcRef,
) {
let mut stack = self.dec_ref_stack.take().unwrap();
debug_assert!(stack.is_empty());
stack.push(gc_ref.unchecked_copy());
while let Some(gc_ref) = stack.pop() {
if self.dec_ref(&gc_ref) {
self.trace_gc_ref(&gc_ref, &mut stack);
if let Some(externref) = gc_ref.as_typed::<VMDrcExternRef>(self) {
let host_data_id = self.index(externref).host_data;
host_data_table.dealloc(host_data_id);
}
self.dealloc(gc_ref.unchecked_copy());
}
}
debug_assert!(stack.is_empty());
debug_assert!(self.dec_ref_stack.is_none());
self.dec_ref_stack = Some(stack);
}
fn ensure_trace_info(&mut self, ty: VMSharedTypeIndex) {
if self.trace_infos.contains_key(&ty) {
return;
}
self.insert_new_trace_info(ty);
}
fn insert_new_trace_info(&mut self, ty: VMSharedTypeIndex) {
debug_assert!(!self.trace_infos.contains_key(&ty));
let engine = self.engine();
let gc_layout = engine
.signatures()
.layout(ty)
.unwrap_or_else(|| panic!("should have a GC layout for {ty:?}"));
let info = match gc_layout {
GcLayout::Array(l) => {
if l.elems_are_gc_refs {
debug_assert_eq!(l.elem_offset(0), GC_REF_ARRAY_ELEMS_OFFSET,);
}
TraceInfo::Array {
gc_ref_elems: l.elems_are_gc_refs,
}
}
GcLayout::Struct(l) => TraceInfo::Struct {
gc_ref_offsets: l
.fields
.iter()
.filter_map(|f| if f.is_gc_ref { Some(f.offset) } else { None })
.collect(),
},
};
let old_entry = self.trace_infos.insert(ty, info);
debug_assert!(old_entry.is_none());
}
fn trace_gc_ref(&self, gc_ref: &VMGcRef, stack: &mut Vec<VMGcRef>) {
debug_assert!(!gc_ref.is_i31());
let header = self.header(gc_ref);
let Some(ty) = header.ty() else {
debug_assert!(header.kind().matches(VMGcKind::ExternRef));
return;
};
match self
.trace_infos
.get(&ty)
.expect("should have inserted trace info for every GC type allocated in this heap")
{
TraceInfo::Struct { gc_ref_offsets } => {
stack.reserve(gc_ref_offsets.len());
let data = self.gc_object_data(gc_ref);
for offset in gc_ref_offsets {
let raw = data.read_u32(*offset);
if let Some(gc_ref) = VMGcRef::from_raw_u32(raw)
&& !gc_ref.is_i31()
{
debug_assert!(
{
let header = self.header(&gc_ref);
let kind = header.kind().as_u32();
VMGcKind::try_from_u32(kind).is_some()
},
"trace_gc_ref: struct field at offset {offset} references object \
with invalid `VMGcKind`",
);
stack.push(gc_ref);
}
}
}
TraceInfo::Array { gc_ref_elems } => {
if !*gc_ref_elems {
return;
}
let data = self.gc_object_data(gc_ref);
let len = self.array_len(gc_ref.as_arrayref_unchecked());
stack.reserve(usize::try_from(len).unwrap());
for i in 0..len {
let elem_offset = GC_REF_ARRAY_ELEMS_OFFSET
+ i * u32::try_from(mem::size_of::<u32>()).unwrap();
let raw = data.read_u32(elem_offset);
if let Some(gc_ref) = VMGcRef::from_raw_u32(raw)
&& !gc_ref.is_i31()
{
debug_assert!(
{
let header = self.header(&gc_ref);
let kind = header.kind().as_u32();
VMGcKind::try_from_u32(kind).is_some()
},
"trace_gc_ref: array element at index {i} references object \
with invalid `VMGcKind`",
);
stack.push(gc_ref);
}
}
}
}
}
fn iter_over_approximated_stack_roots(&self) -> impl Iterator<Item = VMGcRef> + '_ {
let mut link = (*self.over_approximated_stack_roots)
.as_ref()
.map(|r| r.unchecked_copy());
core::iter::from_fn(move || {
let r = link.as_ref()?.unchecked_copy();
link = self.index(drc_ref(&r)).next_over_approximated_stack_root();
Some(r)
})
}
fn assert_over_approximated_stack_roots_integrity(&self) {
if !cfg!(gc_zeal) {
return;
}
let mut visited = HashSet::new();
for gc_ref in self.iter_over_approximated_stack_roots() {
let idx = gc_ref.as_heap_index().unwrap().get();
let header = self.header(&gc_ref);
let kind = header.kind().as_u32();
assert!(
VMGcKind::try_from_u32(kind).is_some(),
"over-approx list: entry at heap index {idx} has invalid VMGcKind {kind:#034b}",
);
let drc_header = self.index(drc_ref(&gc_ref));
assert!(
drc_header.is_in_over_approximated_stack_roots(),
"over-approx list: entry at heap index {idx} does not have in-list bit set",
);
assert_ne!(
drc_header.ref_count, 0,
"over-approx list: entry at heap index {idx} has zero ref count",
);
assert!(
visited.insert(idx),
"over-approx list: cycle or duplicate detected at heap index {idx}",
);
}
}
fn assert_free_blocks_are_poisoned(&self) {
if !cfg!(gc_zeal) {
return;
}
let free_list = self.free_list.as_ref().unwrap();
for (index, len) in free_list.iter_free_blocks() {
let start = usize::try_from(index).unwrap();
let size = usize::try_from(len).unwrap();
let slice = &self.heap_slice()[start..][..size];
assert!(
slice.iter().all(|&b| b == POISON),
"free block at heap index {start} (size {size}) is not fully poisoned",
);
}
}
fn trace(&mut self, roots: &mut GcRootsIter<'_>) {
let mut over_approx_set: DebugOnly<HashSet<_>> = Default::default();
if cfg!(debug_assertions) {
over_approx_set.extend(self.iter_over_approximated_stack_roots());
}
for root in roots {
if !root.is_on_wasm_stack() {
continue;
}
let gc_ref = root.get();
if gc_ref.is_i31() {
continue;
}
log::trace!("Found GC reference on the stack: {gc_ref:#p}");
debug_assert!(
over_approx_set.contains(&gc_ref),
"every on-stack gc ref inside a Wasm frame should \
have be in our over-approximated stack roots set, \
but {gc_ref:#p} is not in the set",
);
debug_assert!(
self.index(drc_ref(&gc_ref))
.is_in_over_approximated_stack_roots(),
"every on-stack gc ref inside a Wasm frame should have \
its in-the-over-approximated-stack-roots-list bit set",
);
debug_assert_ne!(
self.index_mut(drc_ref(&gc_ref)).ref_count,
0,
"{gc_ref:#p} is on the Wasm stack and therefore should be held \
alive by the over-approximated-stack-roots set; should have \
nonzero ref count",
);
self.index_mut(drc_ref(&gc_ref)).set_marked();
}
}
#[inline(never)]
#[cold]
fn log_gc_ref_set(prefix: &str, items: impl Iterator<Item = VMGcRef>) {
assert!(log::log_enabled!(log::Level::Trace));
let mut set = "{".to_string();
let mut any = false;
for gc_ref in items {
any = true;
set += &format!("\n {gc_ref:#p},");
}
if any {
set.push('\n');
}
set.push('}');
log::trace!("{prefix}: {set}");
}
fn sweep(&mut self, host_data_table: &mut ExternRefHostDataTable) {
if log::log_enabled!(log::Level::Trace) {
Self::log_gc_ref_set(
"over-approximated-stack-roots set before sweeping",
self.iter_over_approximated_stack_roots(),
);
}
log::trace!("Begin sweeping");
let mut prev = None;
let mut next = (*self.over_approximated_stack_roots)
.as_ref()
.map(|r| r.unchecked_copy());
while let Some(gc_ref) = next {
log::trace!("sweeping gc ref: {gc_ref:#p}");
let header = self.index_mut(drc_ref(&gc_ref));
debug_assert!(header.is_in_over_approximated_stack_roots());
if header.clear_marked() {
log::trace!(
" -> {gc_ref:#p} is marked, leaving it in the over-approximated-\
stack-roots list"
);
next = header.next_over_approximated_stack_root();
prev = Some(gc_ref);
continue;
}
log::trace!(
" -> {gc_ref:#p} is not marked, removing it from over-approximated-\
stack-roots list and decrementing its ref count"
);
next = header.next_over_approximated_stack_root();
let prev_next = header.next_over_approximated_stack_root();
header.set_in_over_approximated_stack_roots_bit(false);
match &prev {
None => *self.over_approximated_stack_roots = prev_next,
Some(prev) => self
.index_mut(drc_ref(prev))
.set_next_over_approximated_stack_root(prev_next),
}
self.dec_ref_and_maybe_dealloc(host_data_table, &gc_ref);
}
log::trace!("Done sweeping");
if log::log_enabled!(log::Level::Trace) {
Self::log_gc_ref_set(
"over-approximated-stack-roots set after sweeping",
self.iter_over_approximated_stack_roots(),
);
}
}
}
fn drc_ref(gc_ref: &VMGcRef) -> &TypedGcRef<VMDrcHeader> {
debug_assert!(!gc_ref.is_i31());
gc_ref.as_typed_unchecked()
}
fn externref_to_drc(externref: &VMExternRef) -> &TypedGcRef<VMDrcExternRef> {
let gc_ref = externref.as_gc_ref();
debug_assert!(!gc_ref.is_i31());
gc_ref.as_typed_unchecked()
}
#[repr(C)]
struct VMDrcHeader {
header: VMGcHeader,
ref_count: u64,
next_over_approximated_stack_root: Option<VMGcRef>,
object_size: u32,
}
unsafe impl GcHeapObject for VMDrcHeader {
#[inline]
fn is(_header: &VMGcHeader) -> bool {
true
}
}
impl VMDrcHeader {
#[inline]
fn object_size(&self) -> usize {
usize::try_from(self.object_size).unwrap()
}
#[inline]
fn is_in_over_approximated_stack_roots(&self) -> bool {
self.header.reserved_u26() & wasmtime_environ::drc::HEADER_IN_OVER_APPROX_LIST_BIT != 0
}
#[inline]
fn set_in_over_approximated_stack_roots_bit(&mut self, bit: bool) {
let reserved = self.header.reserved_u26();
let new_reserved = if bit {
reserved | wasmtime_environ::drc::HEADER_IN_OVER_APPROX_LIST_BIT
} else {
reserved & !wasmtime_environ::drc::HEADER_IN_OVER_APPROX_LIST_BIT
};
self.header.set_reserved_u26(new_reserved);
}
#[inline]
fn next_over_approximated_stack_root(&self) -> Option<VMGcRef> {
debug_assert!(self.is_in_over_approximated_stack_roots());
self.next_over_approximated_stack_root
.as_ref()
.map(|r| r.unchecked_copy())
}
#[inline]
fn set_next_over_approximated_stack_root(&mut self, next: Option<VMGcRef>) {
debug_assert!(self.is_in_over_approximated_stack_roots());
self.next_over_approximated_stack_root = next;
}
#[inline]
fn is_marked(&self) -> bool {
self.header.reserved_u26() & wasmtime_environ::drc::HEADER_MARK_BIT != 0
}
#[inline]
fn set_marked(&mut self) {
let reserved = self.header.reserved_u26();
self.header
.set_reserved_u26(reserved | wasmtime_environ::drc::HEADER_MARK_BIT);
}
#[inline]
fn clear_marked(&mut self) -> bool {
if self.is_marked() {
let reserved = self.header.reserved_u26();
self.header
.set_reserved_u26(reserved & !wasmtime_environ::drc::HEADER_MARK_BIT);
debug_assert!(!self.is_marked());
true
} else {
false
}
}
}
#[repr(C)]
struct VMDrcArrayHeader {
header: VMDrcHeader,
length: u32,
}
unsafe impl GcHeapObject for VMDrcArrayHeader {
#[inline]
fn is(header: &VMGcHeader) -> bool {
header.kind() == VMGcKind::ArrayRef
}
}
#[repr(C)]
struct VMDrcExternRef {
header: VMDrcHeader,
host_data: ExternRefHostDataId,
}
unsafe impl GcHeapObject for VMDrcExternRef {
#[inline]
fn is(header: &VMGcHeader) -> bool {
header.kind() == VMGcKind::ExternRef
}
}
unsafe impl GcHeap for DrcHeap {
fn is_attached(&self) -> bool {
debug_assert_eq!(self.memory.is_some(), self.free_list.is_some());
debug_assert_eq!(self.memory.is_some(), self.vmmemory.is_some());
self.memory.is_some()
}
fn attach(&mut self, memory: crate::vm::Memory) {
assert!(!self.is_attached());
assert!(!memory.is_shared_memory());
debug_assert!(self.over_approximated_stack_roots.is_none());
let len = memory.vmmemory().current_length();
self.free_list = Some(FreeList::new(len));
self.vmmemory = Some(memory.vmmemory());
self.memory = Some(memory);
if cfg!(gc_zeal) {
self.heap_slice_mut().fill(POISON);
}
}
fn detach(&mut self) -> crate::vm::Memory {
assert!(self.is_attached());
let DrcHeap {
engine: _,
no_gc_count,
over_approximated_stack_roots,
free_list,
dec_ref_stack,
memory,
vmmemory,
trace_infos: _,
} = self;
*no_gc_count = 0;
**over_approximated_stack_roots = None;
*free_list = None;
*vmmemory = None;
debug_assert!(dec_ref_stack.as_ref().is_some_and(|s| s.is_empty()));
memory.take().unwrap()
}
fn as_any(&self) -> &dyn Any {
self as _
}
fn as_any_mut(&mut self) -> &mut dyn Any {
self as _
}
fn enter_no_gc_scope(&mut self) {
self.no_gc_count += 1;
}
fn exit_no_gc_scope(&mut self) {
self.no_gc_count -= 1;
}
fn clone_gc_ref(&mut self, gc_ref: &VMGcRef) -> VMGcRef {
self.inc_ref(gc_ref);
gc_ref.unchecked_copy()
}
fn write_gc_ref(
&mut self,
host_data_table: &mut ExternRefHostDataTable,
destination: &mut Option<VMGcRef>,
source: Option<&VMGcRef>,
) {
if let Some(src) = source {
self.inc_ref(src);
}
if let Some(dest) = destination {
self.dec_ref_and_maybe_dealloc(host_data_table, dest);
}
*destination = source.map(|s| s.unchecked_copy());
}
fn expose_gc_ref_to_wasm(&mut self, gc_ref: VMGcRef) {
let header = self.index_mut(drc_ref(&gc_ref));
if header.is_in_over_approximated_stack_roots() {
return;
}
header.set_in_over_approximated_stack_roots_bit(true);
let next = (*self.over_approximated_stack_roots)
.as_ref()
.map(|r| r.unchecked_copy());
self.index_mut(drc_ref(&gc_ref))
.set_next_over_approximated_stack_root(next);
*self.over_approximated_stack_roots = Some(gc_ref);
}
fn alloc_externref(
&mut self,
host_data: ExternRefHostDataId,
) -> Result<Result<VMExternRef, u64>> {
let gc_ref =
match self.alloc_raw(VMGcHeader::externref(), Layout::new::<VMDrcExternRef>())? {
Err(n) => return Ok(Err(n)),
Ok(gc_ref) => gc_ref,
};
self.index_mut::<VMDrcExternRef>(gc_ref.as_typed_unchecked())
.host_data = host_data;
Ok(Ok(gc_ref.into_externref_unchecked()))
}
fn externref_host_data(&self, externref: &VMExternRef) -> ExternRefHostDataId {
let typed_ref = externref_to_drc(externref);
self.index(typed_ref).host_data
}
fn header(&self, gc_ref: &VMGcRef) -> &VMGcHeader {
let header: &VMGcHeader = self.index(gc_ref.as_typed_unchecked());
debug_assert!(
VMGcKind::try_from_u32(header.kind().as_u32()).is_some(),
"header: invalid VMGcKind {:#010x} at gc_ref {gc_ref:#p}",
header.kind().as_u32(),
);
header
}
fn header_mut(&mut self, gc_ref: &VMGcRef) -> &mut VMGcHeader {
let header: &mut VMGcHeader = self.index_mut(gc_ref.as_typed_unchecked());
debug_assert!(
VMGcKind::try_from_u32(header.kind().as_u32()).is_some(),
"header_mut: invalid VMGcKind {:#010x} at gc_ref {gc_ref:#p}",
header.kind().as_u32(),
);
header
}
fn object_size(&self, gc_ref: &VMGcRef) -> usize {
self.index(drc_ref(gc_ref)).object_size()
}
fn alloc_raw(&mut self, header: VMGcHeader, layout: Layout) -> Result<Result<VMGcRef, u64>> {
debug_assert!(layout.size() >= core::mem::size_of::<VMDrcHeader>());
debug_assert!(layout.align() >= core::mem::align_of::<VMDrcHeader>());
debug_assert_eq!(header.reserved_u26(), 0);
if let Some(ty) = header.ty() {
self.ensure_trace_info(ty);
} else {
debug_assert_eq!(header.kind(), VMGcKind::ExternRef);
}
let object_size = u32::try_from(layout.size()).unwrap();
let gc_ref = match self.free_list.as_mut().unwrap().alloc(layout)? {
None => return Ok(Err(u64::try_from(layout.size()).unwrap())),
Some(index) => VMGcRef::from_heap_index(index).unwrap(),
};
if cfg!(gc_zeal) {
let start = usize::try_from(gc_ref.as_heap_index().unwrap().get()).unwrap();
let slice = &self.heap_slice()[start..][..layout.size()];
gc_assert!(
slice.iter().all(|&b| b == POISON),
"newly allocated GC object at index {start} is not fully poisoned; \
freed memory was corrupted",
);
}
*self.index_mut(drc_ref(&gc_ref)) = VMDrcHeader {
header,
ref_count: 1,
next_over_approximated_stack_root: None,
object_size,
};
log::trace!("new object: increment {gc_ref:#p} ref count -> 1");
Ok(Ok(gc_ref))
}
fn alloc_uninit_struct_or_exn(
&mut self,
ty: VMSharedTypeIndex,
layout: &GcStructLayout,
) -> Result<Result<VMGcRef, u64>> {
let kind = if layout.is_exception {
VMGcKind::ExnRef
} else {
VMGcKind::StructRef
};
let gc_ref =
match self.alloc_raw(VMGcHeader::from_kind_and_index(kind, ty), layout.layout())? {
Err(n) => return Ok(Err(n)),
Ok(gc_ref) => gc_ref,
};
Ok(Ok(gc_ref))
}
fn dealloc_uninit_struct_or_exn(&mut self, gcref: VMGcRef) {
self.dealloc(gcref);
}
fn alloc_uninit_array(
&mut self,
ty: VMSharedTypeIndex,
length: u32,
layout: &GcArrayLayout,
) -> Result<Result<VMArrayRef, u64>> {
let gc_ref = match self.alloc_raw(
VMGcHeader::from_kind_and_index(VMGcKind::ArrayRef, ty),
layout.layout(length),
)? {
Err(n) => return Ok(Err(n)),
Ok(gc_ref) => gc_ref,
};
self.index_mut(gc_ref.as_typed_unchecked::<VMDrcArrayHeader>())
.length = length;
Ok(Ok(gc_ref.into_arrayref_unchecked()))
}
fn dealloc_uninit_array(&mut self, arrayref: VMArrayRef) {
self.dealloc(arrayref.into())
}
fn array_len(&self, arrayref: &VMArrayRef) -> u32 {
debug_assert!(arrayref.as_gc_ref().is_typed::<VMDrcArrayHeader>(self));
self.index::<VMDrcArrayHeader>(arrayref.as_gc_ref().as_typed_unchecked())
.length
}
fn gc<'a>(
&'a mut self,
roots: GcRootsIter<'a>,
host_data_table: &'a mut ExternRefHostDataTable,
) -> Box<dyn GarbageCollection<'a> + 'a> {
assert_eq!(self.no_gc_count, 0, "Cannot GC inside a no-GC scope!");
Box::new(DrcCollection {
roots,
host_data_table,
heap: self,
phase: DrcCollectionPhase::Trace,
})
}
unsafe fn vmctx_gc_heap_data(&self) -> NonNull<u8> {
let ptr: NonNull<Option<VMGcRef>> = NonNull::from(&*self.over_approximated_stack_roots);
ptr.cast()
}
fn take_memory(&mut self) -> crate::vm::Memory {
debug_assert!(self.is_attached());
self.vmmemory.take();
self.memory.take().unwrap()
}
unsafe fn replace_memory(&mut self, memory: crate::vm::Memory, delta_bytes_grown: u64) {
debug_assert!(self.memory.is_none());
debug_assert!(!memory.is_shared_memory());
self.vmmemory = Some(memory.vmmemory());
self.memory = Some(memory);
if cfg!(gc_zeal) {
let old_cap = self.free_list.as_ref().unwrap().current_capacity();
let new_bytes = usize::try_from(delta_bytes_grown).unwrap();
let slice = self.heap_slice_mut();
if old_cap + new_bytes <= slice.len() {
slice[old_cap..old_cap + new_bytes].fill(POISON);
}
}
self.free_list
.as_mut()
.unwrap()
.add_capacity(usize::try_from(delta_bytes_grown).unwrap())
}
#[inline]
fn vmmemory(&self) -> VMMemoryDefinition {
debug_assert!(self.is_attached());
debug_assert!(!self.memory.as_ref().unwrap().is_shared_memory());
let vmmemory = self.vmmemory.as_ref().unwrap();
VMMemoryDefinition {
base: vmmemory.base,
current_length: AtomicUsize::new(vmmemory.current_length()),
}
}
}
struct DrcCollection<'a> {
roots: GcRootsIter<'a>,
host_data_table: &'a mut ExternRefHostDataTable,
heap: &'a mut DrcHeap,
phase: DrcCollectionPhase,
}
enum DrcCollectionPhase {
Trace,
Sweep,
Done,
}
impl<'a> GarbageCollection<'a> for DrcCollection<'a> {
fn collect_increment(&mut self) -> GcProgress {
match self.phase {
DrcCollectionPhase::Trace => {
log::trace!("Begin DRC trace");
self.heap.assert_over_approximated_stack_roots_integrity();
self.heap.assert_free_blocks_are_poisoned();
self.heap.trace(&mut self.roots);
self.heap.assert_over_approximated_stack_roots_integrity();
self.heap.assert_free_blocks_are_poisoned();
log::trace!("End DRC trace");
self.phase = DrcCollectionPhase::Sweep;
GcProgress::Continue
}
DrcCollectionPhase::Sweep => {
log::trace!("Begin DRC sweep");
self.heap.assert_over_approximated_stack_roots_integrity();
self.heap.assert_free_blocks_are_poisoned();
self.heap.sweep(self.host_data_table);
self.heap.assert_over_approximated_stack_roots_integrity();
self.heap.assert_free_blocks_are_poisoned();
log::trace!("End DRC sweep");
self.phase = DrcCollectionPhase::Done;
GcProgress::Complete
}
DrcCollectionPhase::Done => GcProgress::Complete,
}
}
}
#[derive(Debug, Default)]
struct DebugOnly<T> {
inner: T,
}
impl<T> Deref for DebugOnly<T> {
type Target = T;
fn deref(&self) -> &T {
if cfg!(debug_assertions) {
&self.inner
} else {
panic!(
"only deref `DebugOnly` when `cfg(debug_assertions)` or \
inside a `debug_assert!(..)`"
)
}
}
}
impl<T> DerefMut for DebugOnly<T> {
fn deref_mut(&mut self) -> &mut T {
if cfg!(debug_assertions) {
&mut self.inner
} else {
panic!(
"only deref `DebugOnly` when `cfg(debug_assertions)` or \
inside a `debug_assert!(..)`"
)
}
}
}
#[cfg(test)]
mod tests {
use super::*;
use wasmtime_environ::HostPtr;
#[test]
fn vm_drc_header_size_align() {
assert_eq!(
(wasmtime_environ::drc::HEADER_SIZE as usize),
core::mem::size_of::<VMDrcHeader>()
);
assert_eq!(
(wasmtime_environ::drc::HEADER_ALIGN as usize),
core::mem::align_of::<VMDrcHeader>()
);
}
#[test]
fn vm_drc_array_header_length_offset() {
assert_eq!(
wasmtime_environ::drc::ARRAY_LENGTH_OFFSET,
u32::try_from(core::mem::offset_of!(VMDrcArrayHeader, length)).unwrap(),
);
}
#[test]
fn ref_count_is_at_correct_offset() {
let extern_data = VMDrcHeader {
header: VMGcHeader::externref(),
ref_count: 0,
next_over_approximated_stack_root: None,
object_size: 0,
};
let extern_data_ptr = &extern_data as *const _;
let ref_count_ptr = &extern_data.ref_count as *const _;
let actual_offset = (ref_count_ptr as usize) - (extern_data_ptr as usize);
let offsets = wasmtime_environ::VMOffsets::from(wasmtime_environ::VMOffsetsFields {
ptr: HostPtr,
num_imported_functions: 0,
num_imported_tables: 0,
num_imported_memories: 0,
num_imported_globals: 0,
num_imported_tags: 0,
num_defined_tables: 0,
num_defined_memories: 0,
num_owned_memories: 0,
num_defined_globals: 0,
num_defined_tags: 0,
num_escaped_funcs: 0,
});
assert_eq!(
offsets.vm_drc_header_ref_count(),
u32::try_from(actual_offset).unwrap(),
);
}
}