* Add `anyhow` dependency to `wasmtime-runtime`. * Revert `get_data` back to `fn`. * Remove `DataInitializer` and box the data in `Module` translation instead. * Improve comments on `MemoryInitialization`. * Remove `MemoryInitialization::OutOfBounds` in favor of proper bulk memory semantics. * Use segmented memory initialization except for when the uffd feature is enabled on Linux. * Validate modules with the allocator after translation. * Updated various functions in the runtime to return `anyhow::Result`. * Use a slice when copying pages instead of `ptr::copy_nonoverlapping`. * Remove unnecessary casts in `OnDemandAllocator::deallocate`. * Better document the `uffd` feature. * Use WebAssembly page-sized pages in the paged initialization. * Remove the stack pool from the uffd handler and simply protect just the guard pages.
1016 lines
35 KiB
Rust
1016 lines
35 KiB
Rust
//! An `Instance` contains all the runtime state used by execution of a
|
|
//! wasm module (except its callstack and register state). An
|
|
//! `InstanceHandle` is a reference-counting handle for an `Instance`.
|
|
|
|
use crate::export::Export;
|
|
use crate::externref::{StackMapRegistry, VMExternRefActivationsTable};
|
|
use crate::memory::{Memory, RuntimeMemoryCreator};
|
|
use crate::table::{Table, TableElement};
|
|
use crate::traphandlers::Trap;
|
|
use crate::vmcontext::{
|
|
VMBuiltinFunctionsArray, VMCallerCheckedAnyfunc, VMContext, VMFunctionImport,
|
|
VMGlobalDefinition, VMGlobalImport, VMInterrupts, VMMemoryDefinition, VMMemoryImport,
|
|
VMSharedSignatureIndex, VMTableDefinition, VMTableImport,
|
|
};
|
|
use crate::{ExportFunction, ExportGlobal, ExportMemory, ExportTable};
|
|
use indexmap::IndexMap;
|
|
use memoffset::offset_of;
|
|
use more_asserts::assert_lt;
|
|
use std::alloc::Layout;
|
|
use std::any::Any;
|
|
use std::cell::RefCell;
|
|
use std::convert::TryFrom;
|
|
use std::ptr::NonNull;
|
|
use std::rc::Rc;
|
|
use std::sync::Arc;
|
|
use std::{mem, ptr, slice};
|
|
use wasmtime_environ::entity::{packed_option::ReservedValue, EntityRef, EntitySet, PrimaryMap};
|
|
use wasmtime_environ::wasm::{
|
|
DataIndex, DefinedGlobalIndex, DefinedMemoryIndex, DefinedTableIndex, ElemIndex, EntityIndex,
|
|
FuncIndex, GlobalIndex, MemoryIndex, TableElementType, TableIndex,
|
|
};
|
|
use wasmtime_environ::{ir, Module, VMOffsets};
|
|
|
|
mod allocator;
|
|
|
|
pub use allocator::*;
|
|
|
|
/// Runtime representation of an instance value, which erases all `Instance`
|
|
/// information since instances are just a collection of values.
|
|
pub type RuntimeInstance = Rc<IndexMap<String, Export>>;
|
|
|
|
/// A WebAssembly instance.
|
|
///
|
|
/// This is repr(C) to ensure that the vmctx field is last.
|
|
#[repr(C)]
|
|
pub(crate) struct Instance {
|
|
/// The `Module` this `Instance` was instantiated from.
|
|
module: Arc<Module>,
|
|
|
|
/// Offsets in the `vmctx` region.
|
|
offsets: VMOffsets,
|
|
|
|
/// WebAssembly linear memory data.
|
|
memories: PrimaryMap<DefinedMemoryIndex, Memory>,
|
|
|
|
/// WebAssembly table data.
|
|
tables: PrimaryMap<DefinedTableIndex, Table>,
|
|
|
|
/// Stores the dropped passive element segments in this instantiation by index.
|
|
/// If the index is present in the set, the segment has been dropped.
|
|
dropped_elements: RefCell<EntitySet<ElemIndex>>,
|
|
|
|
/// Stores the dropped passive data segments in this instantiation by index.
|
|
/// If the index is present in the set, the segment has been dropped.
|
|
dropped_data: RefCell<EntitySet<DataIndex>>,
|
|
|
|
/// Hosts can store arbitrary per-instance information here.
|
|
host_state: Box<dyn Any>,
|
|
|
|
/// Stores guard page faults in memory relating to the instance.
|
|
/// This is used for the pooling allocator with uffd enabled on Linux.
|
|
#[cfg(all(feature = "uffd", target_os = "linux"))]
|
|
guard_page_faults: RefCell<Vec<(*mut u8, usize, unsafe fn(*mut u8, usize) -> bool)>>,
|
|
|
|
/// Additional context used by compiled wasm code. This field is last, and
|
|
/// represents a dynamically-sized array that extends beyond the nominal
|
|
/// end of the struct (similar to a flexible array member).
|
|
vmctx: VMContext,
|
|
}
|
|
|
|
#[allow(clippy::cast_ptr_alignment)]
|
|
impl Instance {
|
|
/// Helper function to access various locations offset from our `*mut
|
|
/// VMContext` object.
|
|
unsafe fn vmctx_plus_offset<T>(&self, offset: u32) -> *mut T {
|
|
(self.vmctx_ptr() as *mut u8)
|
|
.add(usize::try_from(offset).unwrap())
|
|
.cast()
|
|
}
|
|
|
|
pub(crate) fn module(&self) -> &Module {
|
|
&self.module
|
|
}
|
|
|
|
/// Return a pointer to the `VMSharedSignatureIndex`s.
|
|
fn signature_ids_ptr(&self) -> *mut VMSharedSignatureIndex {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_signature_ids_begin()) }
|
|
}
|
|
|
|
/// Return the indexed `VMFunctionImport`.
|
|
fn imported_function(&self, index: FuncIndex) -> &VMFunctionImport {
|
|
let index = usize::try_from(index.as_u32()).unwrap();
|
|
unsafe { &*self.imported_functions_ptr().add(index) }
|
|
}
|
|
|
|
/// Return a pointer to the `VMFunctionImport`s.
|
|
fn imported_functions_ptr(&self) -> *mut VMFunctionImport {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_imported_functions_begin()) }
|
|
}
|
|
|
|
/// Return the index `VMTableImport`.
|
|
fn imported_table(&self, index: TableIndex) -> &VMTableImport {
|
|
let index = usize::try_from(index.as_u32()).unwrap();
|
|
unsafe { &*self.imported_tables_ptr().add(index) }
|
|
}
|
|
|
|
/// Return a pointer to the `VMTableImports`s.
|
|
fn imported_tables_ptr(&self) -> *mut VMTableImport {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_imported_tables_begin()) }
|
|
}
|
|
|
|
/// Return the indexed `VMMemoryImport`.
|
|
fn imported_memory(&self, index: MemoryIndex) -> &VMMemoryImport {
|
|
let index = usize::try_from(index.as_u32()).unwrap();
|
|
unsafe { &*self.imported_memories_ptr().add(index) }
|
|
}
|
|
|
|
/// Return a pointer to the `VMMemoryImport`s.
|
|
fn imported_memories_ptr(&self) -> *mut VMMemoryImport {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_imported_memories_begin()) }
|
|
}
|
|
|
|
/// Return the indexed `VMGlobalImport`.
|
|
fn imported_global(&self, index: GlobalIndex) -> &VMGlobalImport {
|
|
let index = usize::try_from(index.as_u32()).unwrap();
|
|
unsafe { &*self.imported_globals_ptr().add(index) }
|
|
}
|
|
|
|
/// Return a pointer to the `VMGlobalImport`s.
|
|
fn imported_globals_ptr(&self) -> *mut VMGlobalImport {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_imported_globals_begin()) }
|
|
}
|
|
|
|
/// Return the indexed `VMTableDefinition`.
|
|
#[allow(dead_code)]
|
|
fn table(&self, index: DefinedTableIndex) -> VMTableDefinition {
|
|
unsafe { *self.table_ptr(index) }
|
|
}
|
|
|
|
/// Updates the value for a defined table to `VMTableDefinition`.
|
|
fn set_table(&self, index: DefinedTableIndex, table: VMTableDefinition) {
|
|
unsafe {
|
|
*self.table_ptr(index) = table;
|
|
}
|
|
}
|
|
|
|
/// Return the indexed `VMTableDefinition`.
|
|
fn table_ptr(&self, index: DefinedTableIndex) -> *mut VMTableDefinition {
|
|
let index = usize::try_from(index.as_u32()).unwrap();
|
|
unsafe { self.tables_ptr().add(index) }
|
|
}
|
|
|
|
/// Return a pointer to the `VMTableDefinition`s.
|
|
fn tables_ptr(&self) -> *mut VMTableDefinition {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_tables_begin()) }
|
|
}
|
|
|
|
/// Get a locally defined or imported memory.
|
|
pub(crate) fn get_memory(&self, index: MemoryIndex) -> VMMemoryDefinition {
|
|
if let Some(defined_index) = self.module.defined_memory_index(index) {
|
|
self.memory(defined_index)
|
|
} else {
|
|
let import = self.imported_memory(index);
|
|
*unsafe { import.from.as_ref().unwrap() }
|
|
}
|
|
}
|
|
|
|
/// Return the indexed `VMMemoryDefinition`.
|
|
fn memory(&self, index: DefinedMemoryIndex) -> VMMemoryDefinition {
|
|
unsafe { *self.memory_ptr(index) }
|
|
}
|
|
|
|
/// Set the indexed memory to `VMMemoryDefinition`.
|
|
fn set_memory(&self, index: DefinedMemoryIndex, mem: VMMemoryDefinition) {
|
|
unsafe {
|
|
*self.memory_ptr(index) = mem;
|
|
}
|
|
}
|
|
|
|
/// Return the indexed `VMMemoryDefinition`.
|
|
fn memory_ptr(&self, index: DefinedMemoryIndex) -> *mut VMMemoryDefinition {
|
|
let index = usize::try_from(index.as_u32()).unwrap();
|
|
unsafe { self.memories_ptr().add(index) }
|
|
}
|
|
|
|
/// Return a pointer to the `VMMemoryDefinition`s.
|
|
fn memories_ptr(&self) -> *mut VMMemoryDefinition {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_memories_begin()) }
|
|
}
|
|
|
|
/// Return the indexed `VMGlobalDefinition`.
|
|
fn global(&self, index: DefinedGlobalIndex) -> VMGlobalDefinition {
|
|
unsafe { *self.global_ptr(index) }
|
|
}
|
|
|
|
/// Set the indexed global to `VMGlobalDefinition`.
|
|
#[allow(dead_code)]
|
|
fn set_global(&self, index: DefinedGlobalIndex, global: VMGlobalDefinition) {
|
|
unsafe {
|
|
*self.global_ptr(index) = global;
|
|
}
|
|
}
|
|
|
|
/// Return the indexed `VMGlobalDefinition`.
|
|
fn global_ptr(&self, index: DefinedGlobalIndex) -> *mut VMGlobalDefinition {
|
|
let index = usize::try_from(index.as_u32()).unwrap();
|
|
unsafe { self.globals_ptr().add(index) }
|
|
}
|
|
|
|
/// Get a raw pointer to the global at the given index regardless whether it
|
|
/// is defined locally or imported from another module.
|
|
///
|
|
/// Panics if the index is out of bound or is the reserved value.
|
|
pub(crate) fn defined_or_imported_global_ptr(
|
|
&self,
|
|
index: GlobalIndex,
|
|
) -> *mut VMGlobalDefinition {
|
|
if let Some(index) = self.module.defined_global_index(index) {
|
|
self.global_ptr(index)
|
|
} else {
|
|
self.imported_global(index).from
|
|
}
|
|
}
|
|
|
|
/// Return a pointer to the `VMGlobalDefinition`s.
|
|
fn globals_ptr(&self) -> *mut VMGlobalDefinition {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_globals_begin()) }
|
|
}
|
|
|
|
/// Return a pointer to the `VMBuiltinFunctionsArray`.
|
|
fn builtin_functions_ptr(&self) -> *mut VMBuiltinFunctionsArray {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_builtin_functions_begin()) }
|
|
}
|
|
|
|
/// Return a pointer to the interrupts structure
|
|
pub fn interrupts(&self) -> *mut *const VMInterrupts {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_interrupts()) }
|
|
}
|
|
|
|
/// Return a pointer to the `VMExternRefActivationsTable`.
|
|
pub fn externref_activations_table(&self) -> *mut *mut VMExternRefActivationsTable {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_externref_activations_table()) }
|
|
}
|
|
|
|
/// Return a pointer to the `StackMapRegistry`.
|
|
pub fn stack_map_registry(&self) -> *mut *mut StackMapRegistry {
|
|
unsafe { self.vmctx_plus_offset(self.offsets.vmctx_stack_map_registry()) }
|
|
}
|
|
|
|
/// Return a reference to the vmctx used by compiled wasm code.
|
|
pub fn vmctx(&self) -> &VMContext {
|
|
&self.vmctx
|
|
}
|
|
|
|
/// Return a raw pointer to the vmctx used by compiled wasm code.
|
|
pub fn vmctx_ptr(&self) -> *mut VMContext {
|
|
self.vmctx() as *const VMContext as *mut VMContext
|
|
}
|
|
|
|
/// Lookup an export with the given export declaration.
|
|
pub fn lookup_by_declaration(&self, export: &EntityIndex) -> Export {
|
|
match export {
|
|
EntityIndex::Function(index) => {
|
|
let anyfunc = self.get_caller_checked_anyfunc(*index).unwrap();
|
|
let anyfunc =
|
|
NonNull::new(anyfunc as *const VMCallerCheckedAnyfunc as *mut _).unwrap();
|
|
ExportFunction { anyfunc }.into()
|
|
}
|
|
EntityIndex::Table(index) => {
|
|
let (definition, vmctx) =
|
|
if let Some(def_index) = self.module.defined_table_index(*index) {
|
|
(self.table_ptr(def_index), self.vmctx_ptr())
|
|
} else {
|
|
let import = self.imported_table(*index);
|
|
(import.from, import.vmctx)
|
|
};
|
|
ExportTable {
|
|
definition,
|
|
vmctx,
|
|
table: self.module.table_plans[*index].clone(),
|
|
}
|
|
.into()
|
|
}
|
|
EntityIndex::Memory(index) => {
|
|
let (definition, vmctx) =
|
|
if let Some(def_index) = self.module.defined_memory_index(*index) {
|
|
(self.memory_ptr(def_index), self.vmctx_ptr())
|
|
} else {
|
|
let import = self.imported_memory(*index);
|
|
(import.from, import.vmctx)
|
|
};
|
|
ExportMemory {
|
|
definition,
|
|
vmctx,
|
|
memory: self.module.memory_plans[*index].clone(),
|
|
}
|
|
.into()
|
|
}
|
|
EntityIndex::Global(index) => ExportGlobal {
|
|
definition: if let Some(def_index) = self.module.defined_global_index(*index) {
|
|
self.global_ptr(def_index)
|
|
} else {
|
|
self.imported_global(*index).from
|
|
},
|
|
vmctx: self.vmctx_ptr(),
|
|
global: self.module.globals[*index],
|
|
}
|
|
.into(),
|
|
|
|
EntityIndex::Instance(_) | EntityIndex::Module(_) => {
|
|
panic!("can't use this api for modules/instances")
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Return an iterator over the exports of this instance.
|
|
///
|
|
/// Specifically, it provides access to the key-value pairs, where the keys
|
|
/// are export names, and the values are export declarations which can be
|
|
/// resolved `lookup_by_declaration`.
|
|
pub fn exports(&self) -> indexmap::map::Iter<String, EntityIndex> {
|
|
self.module.exports.iter()
|
|
}
|
|
|
|
/// Return a reference to the custom state attached to this instance.
|
|
#[inline]
|
|
pub fn host_state(&self) -> &dyn Any {
|
|
&*self.host_state
|
|
}
|
|
|
|
/// Return the offset from the vmctx pointer to its containing Instance.
|
|
#[inline]
|
|
pub(crate) fn vmctx_offset() -> isize {
|
|
offset_of!(Self, vmctx) as isize
|
|
}
|
|
|
|
/// Return the table index for the given `VMTableDefinition`.
|
|
pub(crate) fn table_index(&self, table: &VMTableDefinition) -> DefinedTableIndex {
|
|
let offsets = &self.offsets;
|
|
let begin = unsafe {
|
|
(&self.vmctx as *const VMContext as *const u8)
|
|
.add(usize::try_from(offsets.vmctx_tables_begin()).unwrap())
|
|
} as *const VMTableDefinition;
|
|
let end: *const VMTableDefinition = table;
|
|
// TODO: Use `offset_from` once it stablizes.
|
|
let index = DefinedTableIndex::new(
|
|
(end as usize - begin as usize) / mem::size_of::<VMTableDefinition>(),
|
|
);
|
|
assert_lt!(index.index(), self.tables.len());
|
|
index
|
|
}
|
|
|
|
/// Return the memory index for the given `VMMemoryDefinition`.
|
|
pub(crate) fn memory_index(&self, memory: &VMMemoryDefinition) -> DefinedMemoryIndex {
|
|
let offsets = &self.offsets;
|
|
let begin = unsafe {
|
|
(&self.vmctx as *const VMContext as *const u8)
|
|
.add(usize::try_from(offsets.vmctx_memories_begin()).unwrap())
|
|
} as *const VMMemoryDefinition;
|
|
let end: *const VMMemoryDefinition = memory;
|
|
// TODO: Use `offset_from` once it stablizes.
|
|
let index = DefinedMemoryIndex::new(
|
|
(end as usize - begin as usize) / mem::size_of::<VMMemoryDefinition>(),
|
|
);
|
|
assert_lt!(index.index(), self.memories.len());
|
|
index
|
|
}
|
|
|
|
/// Grow memory by the specified amount of pages.
|
|
///
|
|
/// Returns `None` if memory can't be grown by the specified amount
|
|
/// of pages.
|
|
pub(crate) fn memory_grow(&self, memory_index: DefinedMemoryIndex, delta: u32) -> Option<u32> {
|
|
// Reset all guard pages before growing any memory
|
|
#[cfg(all(feature = "uffd", target_os = "linux"))]
|
|
self.reset_guard_pages().ok()?;
|
|
|
|
let result = self
|
|
.memories
|
|
.get(memory_index)
|
|
.unwrap_or_else(|| panic!("no memory for index {}", memory_index.index()))
|
|
.grow(delta);
|
|
|
|
// Keep current the VMContext pointers used by compiled wasm code.
|
|
self.set_memory(memory_index, self.memories[memory_index].vmmemory());
|
|
|
|
result
|
|
}
|
|
|
|
/// Grow imported memory by the specified amount of pages.
|
|
///
|
|
/// Returns `None` if memory can't be grown by the specified amount
|
|
/// of pages.
|
|
///
|
|
/// # Safety
|
|
/// This and `imported_memory_size` are currently unsafe because they
|
|
/// dereference the memory import's pointers.
|
|
pub(crate) unsafe fn imported_memory_grow(
|
|
&self,
|
|
memory_index: MemoryIndex,
|
|
delta: u32,
|
|
) -> Option<u32> {
|
|
let import = self.imported_memory(memory_index);
|
|
let foreign_instance = (&*import.vmctx).instance();
|
|
let foreign_memory = &*import.from;
|
|
let foreign_index = foreign_instance.memory_index(foreign_memory);
|
|
|
|
foreign_instance.memory_grow(foreign_index, delta)
|
|
}
|
|
|
|
/// Returns the number of allocated wasm pages.
|
|
pub(crate) fn memory_size(&self, memory_index: DefinedMemoryIndex) -> u32 {
|
|
self.memories
|
|
.get(memory_index)
|
|
.unwrap_or_else(|| panic!("no memory for index {}", memory_index.index()))
|
|
.size()
|
|
}
|
|
|
|
/// Returns the number of allocated wasm pages in an imported memory.
|
|
///
|
|
/// # Safety
|
|
/// This and `imported_memory_grow` are currently unsafe because they
|
|
/// dereference the memory import's pointers.
|
|
pub(crate) unsafe fn imported_memory_size(&self, memory_index: MemoryIndex) -> u32 {
|
|
let import = self.imported_memory(memory_index);
|
|
let foreign_instance = (&mut *import.vmctx).instance();
|
|
let foreign_memory = &mut *import.from;
|
|
let foreign_index = foreign_instance.memory_index(foreign_memory);
|
|
|
|
foreign_instance.memory_size(foreign_index)
|
|
}
|
|
|
|
pub(crate) fn table_element_type(&self, table_index: TableIndex) -> TableElementType {
|
|
let table = self.get_table(table_index);
|
|
table.element_type()
|
|
}
|
|
|
|
/// Grow table by the specified amount of elements, filling them with
|
|
/// `init_value`.
|
|
///
|
|
/// Returns `None` if table can't be grown by the specified amount of
|
|
/// elements, or if `init_value` is the wrong type of table element.
|
|
pub(crate) fn table_grow(
|
|
&self,
|
|
table_index: TableIndex,
|
|
delta: u32,
|
|
init_value: TableElement,
|
|
) -> Option<u32> {
|
|
let (defined_table_index, instance) =
|
|
self.get_defined_table_index_and_instance(table_index);
|
|
instance.defined_table_grow(defined_table_index, delta, init_value)
|
|
}
|
|
|
|
fn defined_table_grow(
|
|
&self,
|
|
table_index: DefinedTableIndex,
|
|
delta: u32,
|
|
init_value: TableElement,
|
|
) -> Option<u32> {
|
|
unsafe {
|
|
let orig_size = self
|
|
.tables
|
|
.get(table_index)
|
|
.unwrap_or_else(|| panic!("no table for index {}", table_index.index()))
|
|
.grow(delta, init_value)?;
|
|
|
|
// Keep the `VMContext` pointers used by compiled Wasm code up to
|
|
// date.
|
|
self.set_table(table_index, self.tables[table_index].vmtable());
|
|
|
|
Some(orig_size)
|
|
}
|
|
}
|
|
|
|
pub(crate) fn defined_table_fill(
|
|
&self,
|
|
table_index: DefinedTableIndex,
|
|
dst: u32,
|
|
val: TableElement,
|
|
len: u32,
|
|
) -> Result<(), Trap> {
|
|
self.tables.get(table_index).unwrap().fill(dst, val, len)
|
|
}
|
|
|
|
// Get table element by index.
|
|
fn table_get(&self, table_index: DefinedTableIndex, index: u32) -> Option<TableElement> {
|
|
self.tables
|
|
.get(table_index)
|
|
.unwrap_or_else(|| panic!("no table for index {}", table_index.index()))
|
|
.get(index)
|
|
}
|
|
|
|
fn table_set(
|
|
&self,
|
|
table_index: DefinedTableIndex,
|
|
index: u32,
|
|
val: TableElement,
|
|
) -> Result<(), ()> {
|
|
self.tables
|
|
.get(table_index)
|
|
.unwrap_or_else(|| panic!("no table for index {}", table_index.index()))
|
|
.set(index, val)
|
|
}
|
|
|
|
fn alloc_layout(&self) -> Layout {
|
|
let size = mem::size_of_val(self)
|
|
.checked_add(usize::try_from(self.offsets.size_of_vmctx()).unwrap())
|
|
.unwrap();
|
|
let align = mem::align_of_val(self);
|
|
Layout::from_size_align(size, align).unwrap()
|
|
}
|
|
|
|
/// Get a `&VMCallerCheckedAnyfunc` for the given `FuncIndex`.
|
|
///
|
|
/// Returns `None` if the index is the reserved index value.
|
|
///
|
|
/// The returned reference is a stable reference that won't be moved and can
|
|
/// be passed into JIT code.
|
|
pub(crate) fn get_caller_checked_anyfunc(
|
|
&self,
|
|
index: FuncIndex,
|
|
) -> Option<&VMCallerCheckedAnyfunc> {
|
|
if index == FuncIndex::reserved_value() {
|
|
return None;
|
|
}
|
|
|
|
Some(unsafe { &*self.anyfunc_ptr(index) })
|
|
}
|
|
|
|
unsafe fn anyfunc_ptr(&self, index: FuncIndex) -> *mut VMCallerCheckedAnyfunc {
|
|
self.vmctx_plus_offset(self.offsets.vmctx_anyfunc(index))
|
|
}
|
|
|
|
/// The `table.init` operation: initializes a portion of a table with a
|
|
/// passive element.
|
|
///
|
|
/// # Errors
|
|
///
|
|
/// Returns a `Trap` error when the range within the table is out of bounds
|
|
/// or the range within the passive element is out of bounds.
|
|
pub(crate) fn table_init(
|
|
&self,
|
|
table_index: TableIndex,
|
|
elem_index: ElemIndex,
|
|
dst: u32,
|
|
src: u32,
|
|
len: u32,
|
|
) -> Result<(), Trap> {
|
|
// https://webassembly.github.io/bulk-memory-operations/core/exec/instructions.html#exec-table-init
|
|
|
|
let table = self.get_table(table_index);
|
|
let elem_index = self.module.passive_elements_map.get(&elem_index);
|
|
let elem = match elem_index {
|
|
Some(index) => {
|
|
if self
|
|
.dropped_elements
|
|
.borrow()
|
|
.contains(ElemIndex::new(*index))
|
|
{
|
|
&[]
|
|
} else {
|
|
self.module.passive_elements[*index].as_ref()
|
|
}
|
|
}
|
|
None => &[],
|
|
};
|
|
|
|
if src
|
|
.checked_add(len)
|
|
.map_or(true, |n| n as usize > elem.len())
|
|
|| dst.checked_add(len).map_or(true, |m| m > table.size())
|
|
{
|
|
return Err(Trap::wasm(ir::TrapCode::TableOutOfBounds));
|
|
}
|
|
|
|
// TODO(#983): investigate replacing this get/set loop with a `memcpy`.
|
|
for (dst, src) in (dst..dst + len).zip(src..src + len) {
|
|
let elem = self
|
|
.get_caller_checked_anyfunc(elem[src as usize])
|
|
.map_or(ptr::null_mut(), |f: &VMCallerCheckedAnyfunc| {
|
|
f as *const VMCallerCheckedAnyfunc as *mut _
|
|
});
|
|
|
|
table
|
|
.set(dst, TableElement::FuncRef(elem))
|
|
.expect("should never panic because we already did the bounds check above");
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Drop an element.
|
|
pub(crate) fn elem_drop(&self, elem_index: ElemIndex) {
|
|
// https://webassembly.github.io/reference-types/core/exec/instructions.html#exec-elem-drop
|
|
|
|
if let Some(index) = self.module.passive_elements_map.get(&elem_index) {
|
|
self.dropped_elements
|
|
.borrow_mut()
|
|
.insert(ElemIndex::new(*index));
|
|
}
|
|
|
|
// Note that we don't check that we actually removed a segment because
|
|
// dropping a non-passive segment is a no-op (not a trap).
|
|
}
|
|
|
|
/// Do a `memory.copy`
|
|
///
|
|
/// # Errors
|
|
///
|
|
/// Returns a `Trap` error when the source or destination ranges are out of
|
|
/// bounds.
|
|
pub(crate) fn memory_copy(
|
|
&self,
|
|
dst_index: MemoryIndex,
|
|
dst: u32,
|
|
src_index: MemoryIndex,
|
|
src: u32,
|
|
len: u32,
|
|
) -> Result<(), Trap> {
|
|
// https://webassembly.github.io/reference-types/core/exec/instructions.html#exec-memory-copy
|
|
|
|
let src_mem = self.get_memory(src_index);
|
|
let dst_mem = self.get_memory(dst_index);
|
|
|
|
if src
|
|
.checked_add(len)
|
|
.map_or(true, |n| n as usize > src_mem.current_length)
|
|
|| dst
|
|
.checked_add(len)
|
|
.map_or(true, |m| m as usize > dst_mem.current_length)
|
|
{
|
|
return Err(Trap::wasm(ir::TrapCode::HeapOutOfBounds));
|
|
}
|
|
|
|
let dst = usize::try_from(dst).unwrap();
|
|
let src = usize::try_from(src).unwrap();
|
|
|
|
// Bounds and casts are checked above, by this point we know that
|
|
// everything is safe.
|
|
unsafe {
|
|
let dst = dst_mem.base.add(dst);
|
|
let src = src_mem.base.add(src);
|
|
ptr::copy(src, dst, len as usize);
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Perform the `memory.fill` operation on a locally defined memory.
|
|
///
|
|
/// # Errors
|
|
///
|
|
/// Returns a `Trap` error if the memory range is out of bounds.
|
|
pub(crate) fn defined_memory_fill(
|
|
&self,
|
|
memory_index: DefinedMemoryIndex,
|
|
dst: u32,
|
|
val: u32,
|
|
len: u32,
|
|
) -> Result<(), Trap> {
|
|
let memory = self.memory(memory_index);
|
|
|
|
if dst
|
|
.checked_add(len)
|
|
.map_or(true, |m| m as usize > memory.current_length)
|
|
{
|
|
return Err(Trap::wasm(ir::TrapCode::HeapOutOfBounds));
|
|
}
|
|
|
|
let dst = isize::try_from(dst).unwrap();
|
|
let val = val as u8;
|
|
|
|
// Bounds and casts are checked above, by this point we know that
|
|
// everything is safe.
|
|
unsafe {
|
|
let dst = memory.base.offset(dst);
|
|
ptr::write_bytes(dst, val, len as usize);
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Perform the `memory.fill` operation on an imported memory.
|
|
///
|
|
/// # Errors
|
|
///
|
|
/// Returns a `Trap` error if the memory range is out of bounds.
|
|
pub(crate) fn imported_memory_fill(
|
|
&self,
|
|
memory_index: MemoryIndex,
|
|
dst: u32,
|
|
val: u32,
|
|
len: u32,
|
|
) -> Result<(), Trap> {
|
|
let import = self.imported_memory(memory_index);
|
|
unsafe {
|
|
let foreign_instance = (&*import.vmctx).instance();
|
|
let foreign_memory = &*import.from;
|
|
let foreign_index = foreign_instance.memory_index(foreign_memory);
|
|
foreign_instance.defined_memory_fill(foreign_index, dst, val, len)
|
|
}
|
|
}
|
|
|
|
/// Performs the `memory.init` operation.
|
|
///
|
|
/// # Errors
|
|
///
|
|
/// Returns a `Trap` error if the destination range is out of this module's
|
|
/// memory's bounds or if the source range is outside the data segment's
|
|
/// bounds.
|
|
pub(crate) fn memory_init(
|
|
&self,
|
|
memory_index: MemoryIndex,
|
|
data_index: DataIndex,
|
|
dst: u32,
|
|
src: u32,
|
|
len: u32,
|
|
) -> Result<(), Trap> {
|
|
// https://webassembly.github.io/bulk-memory-operations/core/exec/instructions.html#exec-memory-init
|
|
|
|
let memory = self.get_memory(memory_index);
|
|
let data_index = self.module.passive_data_map.get(&data_index);
|
|
let data = match data_index {
|
|
Some(index) => {
|
|
if self.dropped_data.borrow().contains(DataIndex::new(*index)) {
|
|
&[]
|
|
} else {
|
|
self.module.passive_data[*index].as_ref()
|
|
}
|
|
}
|
|
None => &[],
|
|
};
|
|
|
|
if src
|
|
.checked_add(len)
|
|
.map_or(true, |n| n as usize > data.len())
|
|
|| dst
|
|
.checked_add(len)
|
|
.map_or(true, |m| m as usize > memory.current_length)
|
|
{
|
|
return Err(Trap::wasm(ir::TrapCode::HeapOutOfBounds));
|
|
}
|
|
|
|
let src_slice = &data[src as usize..(src + len) as usize];
|
|
|
|
unsafe {
|
|
let dst_start = memory.base.add(dst as usize);
|
|
let dst_slice = slice::from_raw_parts_mut(dst_start, len as usize);
|
|
dst_slice.copy_from_slice(src_slice);
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Drop the given data segment, truncating its length to zero.
|
|
pub(crate) fn data_drop(&self, data_index: DataIndex) {
|
|
if let Some(index) = self.module.passive_data_map.get(&data_index) {
|
|
self.dropped_data
|
|
.borrow_mut()
|
|
.insert(DataIndex::new(*index));
|
|
}
|
|
|
|
// Note that we don't check that we actually removed a segment because
|
|
// dropping a non-passive segment is a no-op (not a trap).
|
|
}
|
|
|
|
/// Get a table by index regardless of whether it is locally-defined or an
|
|
/// imported, foreign table.
|
|
pub(crate) fn get_table(&self, table_index: TableIndex) -> &Table {
|
|
if let Some(defined_table_index) = self.module.defined_table_index(table_index) {
|
|
self.get_defined_table(defined_table_index)
|
|
} else {
|
|
self.get_foreign_table(table_index)
|
|
}
|
|
}
|
|
|
|
/// Get a locally-defined table.
|
|
pub(crate) fn get_defined_table(&self, index: DefinedTableIndex) -> &Table {
|
|
&self.tables[index]
|
|
}
|
|
|
|
/// Get an imported, foreign table.
|
|
pub(crate) fn get_foreign_table(&self, index: TableIndex) -> &Table {
|
|
let import = self.imported_table(index);
|
|
let foreign_instance = unsafe { (&mut *(import).vmctx).instance() };
|
|
let foreign_table = unsafe { &mut *(import).from };
|
|
let foreign_index = foreign_instance.table_index(foreign_table);
|
|
&foreign_instance.tables[foreign_index]
|
|
}
|
|
|
|
pub(crate) fn get_defined_table_index_and_instance(
|
|
&self,
|
|
index: TableIndex,
|
|
) -> (DefinedTableIndex, &Instance) {
|
|
if let Some(defined_table_index) = self.module.defined_table_index(index) {
|
|
(defined_table_index, self)
|
|
} else {
|
|
let import = self.imported_table(index);
|
|
let foreign_instance = unsafe { (&mut *(import).vmctx).instance() };
|
|
let foreign_table_def = unsafe { &mut *(import).from };
|
|
let foreign_table_index = foreign_instance.table_index(foreign_table_def);
|
|
(foreign_table_index, foreign_instance)
|
|
}
|
|
}
|
|
|
|
/// Records a faulted guard page.
|
|
///
|
|
/// This is used to track faulted guard pages that need to be reset.
|
|
#[cfg(all(feature = "uffd", target_os = "linux"))]
|
|
pub(crate) fn record_guard_page_fault(
|
|
&self,
|
|
page_addr: *mut u8,
|
|
size: usize,
|
|
reset: unsafe fn(*mut u8, usize) -> bool,
|
|
) {
|
|
self.guard_page_faults
|
|
.borrow_mut()
|
|
.push((page_addr, size, reset));
|
|
}
|
|
|
|
/// Resets previously faulted guard pages.
|
|
///
|
|
/// This is used to reset the protection of any guard pages that were previously faulted.
|
|
///
|
|
/// Resetting the guard pages is required before growing memory.
|
|
#[cfg(all(feature = "uffd", target_os = "linux"))]
|
|
pub(crate) fn reset_guard_pages(&self) -> anyhow::Result<()> {
|
|
let mut faults = self.guard_page_faults.borrow_mut();
|
|
for (addr, len, reset) in faults.drain(..) {
|
|
unsafe {
|
|
if !reset(addr, len) {
|
|
anyhow::bail!("failed to reset previously faulted memory guard page");
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
/// A handle holding an `Instance` of a WebAssembly module.
|
|
#[derive(Hash, PartialEq, Eq)]
|
|
pub struct InstanceHandle {
|
|
instance: *mut Instance,
|
|
}
|
|
|
|
impl InstanceHandle {
|
|
pub(crate) unsafe fn new(instance: *mut Instance) -> Self {
|
|
Self { instance }
|
|
}
|
|
|
|
/// Create a new `InstanceHandle` pointing at the instance
|
|
/// pointed to by the given `VMContext` pointer.
|
|
///
|
|
/// # Safety
|
|
/// This is unsafe because it doesn't work on just any `VMContext`, it must
|
|
/// be a `VMContext` allocated as part of an `Instance`.
|
|
pub unsafe fn from_vmctx(vmctx: *mut VMContext) -> Self {
|
|
let instance = (&mut *vmctx).instance();
|
|
Self {
|
|
instance: instance as *const Instance as *mut Instance,
|
|
}
|
|
}
|
|
|
|
/// Return a reference to the vmctx used by compiled wasm code.
|
|
pub fn vmctx(&self) -> &VMContext {
|
|
self.instance().vmctx()
|
|
}
|
|
|
|
/// Return a raw pointer to the vmctx used by compiled wasm code.
|
|
pub fn vmctx_ptr(&self) -> *mut VMContext {
|
|
self.instance().vmctx_ptr()
|
|
}
|
|
|
|
/// Return a reference to a module.
|
|
pub fn module(&self) -> &Module {
|
|
self.instance().module()
|
|
}
|
|
|
|
/// Lookup an export with the given export declaration.
|
|
pub fn lookup_by_declaration(&self, export: &EntityIndex) -> Export {
|
|
self.instance().lookup_by_declaration(export)
|
|
}
|
|
|
|
/// Return an iterator over the exports of this instance.
|
|
///
|
|
/// Specifically, it provides access to the key-value pairs, where the keys
|
|
/// are export names, and the values are export declarations which can be
|
|
/// resolved `lookup_by_declaration`.
|
|
pub fn exports(&self) -> indexmap::map::Iter<String, EntityIndex> {
|
|
self.instance().exports()
|
|
}
|
|
|
|
/// Return a reference to the custom state attached to this instance.
|
|
pub fn host_state(&self) -> &dyn Any {
|
|
self.instance().host_state()
|
|
}
|
|
|
|
/// Return the memory index for the given `VMMemoryDefinition` in this instance.
|
|
pub fn memory_index(&self, memory: &VMMemoryDefinition) -> DefinedMemoryIndex {
|
|
self.instance().memory_index(memory)
|
|
}
|
|
|
|
/// Grow memory in this instance by the specified amount of pages.
|
|
///
|
|
/// Returns `None` if memory can't be grown by the specified amount
|
|
/// of pages.
|
|
pub fn memory_grow(&self, memory_index: DefinedMemoryIndex, delta: u32) -> Option<u32> {
|
|
self.instance().memory_grow(memory_index, delta)
|
|
}
|
|
|
|
/// Return the table index for the given `VMTableDefinition` in this instance.
|
|
pub fn table_index(&self, table: &VMTableDefinition) -> DefinedTableIndex {
|
|
self.instance().table_index(table)
|
|
}
|
|
|
|
/// Grow table in this instance by the specified amount of elements.
|
|
///
|
|
/// When the table is successfully grown, returns the original size of the
|
|
/// table.
|
|
///
|
|
/// Returns `None` if memory can't be grown by the specified amount of pages
|
|
/// or if the `init_value` is the incorrect table element type.
|
|
pub fn table_grow(
|
|
&self,
|
|
table_index: TableIndex,
|
|
delta: u32,
|
|
init_value: TableElement,
|
|
) -> Option<u32> {
|
|
self.instance().table_grow(table_index, delta, init_value)
|
|
}
|
|
|
|
/// Grow table in this instance by the specified amount of elements.
|
|
///
|
|
/// When the table is successfully grown, returns the original size of the
|
|
/// table.
|
|
///
|
|
/// Returns `None` if memory can't be grown by the specified amount of pages
|
|
/// or if the `init_value` is the incorrect table element type.
|
|
pub fn defined_table_grow(
|
|
&self,
|
|
table_index: DefinedTableIndex,
|
|
delta: u32,
|
|
init_value: TableElement,
|
|
) -> Option<u32> {
|
|
self.instance()
|
|
.defined_table_grow(table_index, delta, init_value)
|
|
}
|
|
|
|
/// Get table element reference.
|
|
///
|
|
/// Returns `None` if index is out of bounds.
|
|
pub fn table_get(&self, table_index: DefinedTableIndex, index: u32) -> Option<TableElement> {
|
|
self.instance().table_get(table_index, index)
|
|
}
|
|
|
|
/// Set table element reference.
|
|
///
|
|
/// Returns an error if the index is out of bounds
|
|
pub fn table_set(
|
|
&self,
|
|
table_index: DefinedTableIndex,
|
|
index: u32,
|
|
val: TableElement,
|
|
) -> Result<(), ()> {
|
|
self.instance().table_set(table_index, index, val)
|
|
}
|
|
|
|
/// Fill a region of the table.
|
|
///
|
|
/// Returns an error if the region is out of bounds or val is not of the
|
|
/// correct type.
|
|
pub fn defined_table_fill(
|
|
&self,
|
|
table_index: DefinedTableIndex,
|
|
dst: u32,
|
|
val: TableElement,
|
|
len: u32,
|
|
) -> Result<(), Trap> {
|
|
self.instance()
|
|
.defined_table_fill(table_index, dst, val, len)
|
|
}
|
|
|
|
/// Get a table defined locally within this module.
|
|
pub fn get_defined_table(&self, index: DefinedTableIndex) -> &Table {
|
|
self.instance().get_defined_table(index)
|
|
}
|
|
|
|
/// Return a reference to the contained `Instance`.
|
|
pub(crate) fn instance(&self) -> &Instance {
|
|
unsafe { &*(self.instance as *const Instance) }
|
|
}
|
|
|
|
/// Returns a clone of this instance.
|
|
///
|
|
/// This is unsafe because the returned handle here is just a cheap clone
|
|
/// of the internals, there's no lifetime tracking around its validity.
|
|
/// You'll need to ensure that the returned handles all go out of scope at
|
|
/// the same time.
|
|
pub unsafe fn clone(&self) -> InstanceHandle {
|
|
InstanceHandle {
|
|
instance: self.instance,
|
|
}
|
|
}
|
|
}
|