Files
wasmtime/cranelift/interpreter/src/interpreter.rs
Nick Fitzgerald c0b587ac5f Remove heaps from core Cranelift, push them into cranelift-wasm (#5386)
* cranelift-wasm: translate Wasm loads into lower-level CLIF operations

Rather than using `heap_{load,store,addr}`.

* cranelift: Remove the `heap_{addr,load,store}` instructions

These are now legalized in the `cranelift-wasm` frontend.

* cranelift: Remove the `ir::Heap` entity from CLIF

* Port basic memory operation tests to .wat filetests

* Remove test for verifying CLIF heaps

* Remove `heap_addr` from replace_branching_instructions_and_cfg_predecessors.clif test

* Remove `heap_addr` from readonly.clif test

* Remove `heap_addr` from `table_addr.clif` test

* Remove `heap_addr` from the simd-fvpromote_low.clif test

* Remove `heap_addr` from simd-fvdemote.clif test

* Remove `heap_addr` from the load-op-store.clif test

* Remove the CLIF heap runtest

* Remove `heap_addr` from the global_value.clif test

* Remove `heap_addr` from fpromote.clif runtests

* Remove `heap_addr` from fdemote.clif runtests

* Remove `heap_addr` from memory.clif parser test

* Remove `heap_addr` from reject_load_readonly.clif test

* Remove `heap_addr` from reject_load_notrap.clif test

* Remove `heap_addr` from load_readonly_notrap.clif test

* Remove `static-heap-without-guard-pages.clif` test

Will be subsumed when we port `make-heap-load-store-tests.sh` to generating
`.wat` tests.

* Remove `static-heap-with-guard-pages.clif` test

Will be subsumed when we port `make-heap-load-store-tests.sh` over to `.wat`
tests.

* Remove more heap tests

These will be subsumed by porting `make-heap-load-store-tests.sh` over to `.wat`
tests.

* Remove `heap_addr` from `simple-alias.clif` test

* Remove `heap_addr` from partial-redundancy.clif test

* Remove `heap_addr` from multiple-blocks.clif test

* Remove `heap_addr` from fence.clif test

* Remove `heap_addr` from extends.clif test

* Remove runtests that rely on heaps

Heaps are not a thing in CLIF or the interpreter anymore

* Add generated load/store `.wat` tests

* Enable memory-related wasm features in `.wat` tests

* Remove CLIF heap from fcmp-mem-bug.clif test

* Add a mode for compiling `.wat` all the way to assembly in filetests

* Also generate WAT to assembly tests in `make-load-store-tests.sh`

* cargo fmt

* Reinstate `f{de,pro}mote.clif` tests without the heap bits

* Remove undefined doc link

* Remove outdated SVG and dot file from docs

* Add docs about `None` returns for base address computation helpers

* Factor out `env.heap_access_spectre_mitigation()` to a local

* Expand docs for `FuncEnvironment::heaps` trait method

* Restore f{de,pro}mote+load clif runtests with stack memory
2022-12-15 00:26:45 +00:00

883 lines
30 KiB
Rust

//! Cranelift IR interpreter.
//!
//! This module partially contains the logic for interpreting Cranelift IR.
use crate::address::{Address, AddressRegion, AddressSize};
use crate::environment::{FuncIndex, FunctionStore};
use crate::frame::Frame;
use crate::instruction::DfgInstructionContext;
use crate::state::{MemoryError, State};
use crate::step::{step, ControlFlow, StepError};
use crate::value::{Value, ValueError};
use cranelift_codegen::data_value::DataValue;
use cranelift_codegen::ir::{
ArgumentPurpose, Block, FuncRef, Function, GlobalValue, GlobalValueData, LibCall, StackSlot,
TrapCode, Type, Value as ValueRef,
};
use log::trace;
use smallvec::SmallVec;
use std::convert::TryFrom;
use std::fmt::Debug;
use std::iter;
use thiserror::Error;
/// The Cranelift interpreter; this contains some high-level functions to control the interpreter's
/// flow. The interpreter state is defined separately (see [InterpreterState]) as the execution
/// semantics for each Cranelift instruction (see [step]).
pub struct Interpreter<'a> {
state: InterpreterState<'a>,
fuel: Option<u64>,
}
impl<'a> Interpreter<'a> {
pub fn new(state: InterpreterState<'a>) -> Self {
Self { state, fuel: None }
}
/// The `fuel` mechanism sets a number of instructions that
/// the interpreter can execute before stopping. If this
/// value is `None` (the default), no limit is imposed.
pub fn with_fuel(self, fuel: Option<u64>) -> Self {
Self { fuel, ..self }
}
/// Call a function by name; this is a helpful proxy for [Interpreter::call_by_index].
pub fn call_by_name(
&mut self,
func_name: &str,
arguments: &[DataValue],
) -> Result<ControlFlow<'a, DataValue>, InterpreterError> {
let index = self
.state
.functions
.index_of(func_name)
.ok_or_else(|| InterpreterError::UnknownFunctionName(func_name.to_string()))?;
self.call_by_index(index, arguments)
}
/// Call a function by its index in the [FunctionStore]; this is a proxy for
/// `Interpreter::call`.
pub fn call_by_index(
&mut self,
index: FuncIndex,
arguments: &[DataValue],
) -> Result<ControlFlow<'a, DataValue>, InterpreterError> {
match self.state.functions.get_by_index(index) {
None => Err(InterpreterError::UnknownFunctionIndex(index)),
Some(func) => self.call(func, arguments),
}
}
/// Interpret a call to a [Function] given its [DataValue] arguments.
fn call(
&mut self,
function: &'a Function,
arguments: &[DataValue],
) -> Result<ControlFlow<'a, DataValue>, InterpreterError> {
trace!("Call: {}({:?})", function.name, arguments);
let first_block = function
.layout
.blocks()
.next()
.expect("to have a first block");
let parameters = function.dfg.block_params(first_block);
self.state.push_frame(function);
self.state
.current_frame_mut()
.set_all(parameters, arguments.to_vec());
self.block(first_block)
}
/// Interpret a [Block] in a [Function]. This drives the interpretation over sequences of
/// instructions, which may continue in other blocks, until the function returns.
fn block(&mut self, block: Block) -> Result<ControlFlow<'a, DataValue>, InterpreterError> {
trace!("Block: {}", block);
let function = self.state.current_frame_mut().function;
let layout = &function.layout;
let mut maybe_inst = layout.first_inst(block);
while let Some(inst) = maybe_inst {
if self.consume_fuel() == FuelResult::Stop {
return Err(InterpreterError::FuelExhausted);
}
let inst_context = DfgInstructionContext::new(inst, &function.dfg);
match step(&mut self.state, inst_context)? {
ControlFlow::Assign(values) => {
self.state
.current_frame_mut()
.set_all(function.dfg.inst_results(inst), values.to_vec());
maybe_inst = layout.next_inst(inst)
}
ControlFlow::Continue => maybe_inst = layout.next_inst(inst),
ControlFlow::ContinueAt(block, block_arguments) => {
trace!("Block: {}", block);
self.state
.current_frame_mut()
.set_all(function.dfg.block_params(block), block_arguments.to_vec());
maybe_inst = layout.first_inst(block)
}
ControlFlow::Call(called_function, arguments) => {
let returned_arguments =
self.call(called_function, &arguments)?.unwrap_return();
self.state
.current_frame_mut()
.set_all(function.dfg.inst_results(inst), returned_arguments);
maybe_inst = layout.next_inst(inst)
}
ControlFlow::Return(returned_values) => {
self.state.pop_frame();
return Ok(ControlFlow::Return(returned_values));
}
ControlFlow::Trap(trap) => return Ok(ControlFlow::Trap(trap)),
}
}
Err(InterpreterError::Unreachable)
}
fn consume_fuel(&mut self) -> FuelResult {
match self.fuel {
Some(0) => FuelResult::Stop,
Some(ref mut n) => {
*n -= 1;
FuelResult::Continue
}
// We do not have fuel enabled, so unconditionally continue
None => FuelResult::Continue,
}
}
}
#[derive(Debug, PartialEq, Clone)]
/// The result of consuming fuel. Signals if the caller should stop or continue.
pub enum FuelResult {
/// We still have `fuel` available and should continue execution.
Continue,
/// The available `fuel` has been exhausted, we should stop now.
Stop,
}
/// The ways interpretation can fail.
#[derive(Error, Debug)]
pub enum InterpreterError {
#[error("failed to interpret instruction")]
StepError(#[from] StepError),
#[error("reached an unreachable statement")]
Unreachable,
#[error("unknown function index (has it been added to the function store?): {0}")]
UnknownFunctionIndex(FuncIndex),
#[error("unknown function with name (has it been added to the function store?): {0}")]
UnknownFunctionName(String),
#[error("value error")]
ValueError(#[from] ValueError),
#[error("fuel exhausted")]
FuelExhausted,
}
pub type LibCallValues<V> = SmallVec<[V; 1]>;
pub type LibCallHandler<V> = fn(LibCall, LibCallValues<V>) -> Result<LibCallValues<V>, TrapCode>;
/// Maintains the [Interpreter]'s state, implementing the [State] trait.
pub struct InterpreterState<'a> {
pub functions: FunctionStore<'a>,
pub libcall_handler: LibCallHandler<DataValue>,
pub frame_stack: Vec<Frame<'a>>,
/// Number of bytes from the bottom of the stack where the current frame's stack space is
pub frame_offset: usize,
pub stack: Vec<u8>,
pub pinned_reg: DataValue,
}
impl Default for InterpreterState<'_> {
fn default() -> Self {
Self {
functions: FunctionStore::default(),
libcall_handler: |_, _| Err(TrapCode::UnreachableCodeReached),
frame_stack: vec![],
frame_offset: 0,
stack: Vec::with_capacity(1024),
pinned_reg: DataValue::U64(0),
}
}
}
impl<'a> InterpreterState<'a> {
pub fn with_function_store(self, functions: FunctionStore<'a>) -> Self {
Self { functions, ..self }
}
/// Registers a libcall handler
pub fn with_libcall_handler(mut self, handler: LibCallHandler<DataValue>) -> Self {
self.libcall_handler = handler;
self
}
fn current_frame_mut(&mut self) -> &mut Frame<'a> {
let num_frames = self.frame_stack.len();
match num_frames {
0 => panic!("unable to retrieve the current frame because no frames were pushed"),
_ => &mut self.frame_stack[num_frames - 1],
}
}
fn current_frame(&self) -> &Frame<'a> {
let num_frames = self.frame_stack.len();
match num_frames {
0 => panic!("unable to retrieve the current frame because no frames were pushed"),
_ => &self.frame_stack[num_frames - 1],
}
}
}
impl<'a> State<'a, DataValue> for InterpreterState<'a> {
fn get_function(&self, func_ref: FuncRef) -> Option<&'a Function> {
self.functions
.get_from_func_ref(func_ref, self.frame_stack.last().unwrap().function)
}
fn get_current_function(&self) -> &'a Function {
self.current_frame().function
}
fn get_libcall_handler(&self) -> LibCallHandler<DataValue> {
self.libcall_handler
}
fn push_frame(&mut self, function: &'a Function) {
if let Some(frame) = self.frame_stack.iter().last() {
self.frame_offset += frame.function.fixed_stack_size() as usize;
}
// Grow the stack by the space necessary for this frame
self.stack
.extend(iter::repeat(0).take(function.fixed_stack_size() as usize));
self.frame_stack.push(Frame::new(function));
}
fn pop_frame(&mut self) {
if let Some(frame) = self.frame_stack.pop() {
// Shorten the stack after exiting the frame
self.stack
.truncate(self.stack.len() - frame.function.fixed_stack_size() as usize);
// Reset frame_offset to the start of this function
if let Some(frame) = self.frame_stack.iter().last() {
self.frame_offset -= frame.function.fixed_stack_size() as usize;
}
}
}
fn get_value(&self, name: ValueRef) -> Option<DataValue> {
Some(self.current_frame().get(name).clone()) // TODO avoid clone?
}
fn set_value(&mut self, name: ValueRef, value: DataValue) -> Option<DataValue> {
self.current_frame_mut().set(name, value)
}
fn stack_address(
&self,
size: AddressSize,
slot: StackSlot,
offset: u64,
) -> Result<Address, MemoryError> {
let stack_slots = &self.get_current_function().sized_stack_slots;
let stack_slot = &stack_slots[slot];
// offset must be `0 <= Offset < sizeof(SS)`
if offset >= stack_slot.size as u64 {
return Err(MemoryError::InvalidOffset {
offset,
max: stack_slot.size as u64,
});
}
// Calculate the offset from the current frame to the requested stack slot
let slot_offset: u64 = stack_slots
.keys()
.filter(|k| k < &slot)
.map(|k| stack_slots[k].size as u64)
.sum();
let final_offset = self.frame_offset as u64 + slot_offset + offset;
Address::from_parts(size, AddressRegion::Stack, 0, final_offset)
}
fn checked_load(&self, addr: Address, ty: Type) -> Result<DataValue, MemoryError> {
let load_size = ty.bytes() as usize;
let addr_start = addr.offset as usize;
let addr_end = addr_start + load_size;
let src = match addr.region {
AddressRegion::Stack => {
if addr_end > self.stack.len() {
return Err(MemoryError::OutOfBoundsLoad { addr, load_size });
}
&self.stack[addr_start..addr_end]
}
_ => unimplemented!(),
};
Ok(DataValue::read_from_slice(src, ty))
}
fn checked_store(&mut self, addr: Address, v: DataValue) -> Result<(), MemoryError> {
let store_size = v.ty().bytes() as usize;
let addr_start = addr.offset as usize;
let addr_end = addr_start + store_size;
let dst = match addr.region {
AddressRegion::Stack => {
if addr_end > self.stack.len() {
return Err(MemoryError::OutOfBoundsStore { addr, store_size });
}
&mut self.stack[addr_start..addr_end]
}
_ => unimplemented!(),
};
Ok(v.write_to_slice(dst))
}
/// Non-Recursively resolves a global value until its address is found
fn resolve_global_value(&self, gv: GlobalValue) -> Result<DataValue, MemoryError> {
// Resolving a Global Value is a "pointer" chasing operation that lends itself to
// using a recursive solution. However, resolving this in a recursive manner
// is a bad idea because its very easy to add a bunch of global values and
// blow up the call stack.
//
// Adding to the challenges of this, is that the operations possible with GlobalValues
// mean that we cannot use a simple loop to resolve each global value, we must keep
// a pending list of operations.
// These are the possible actions that we can perform
#[derive(Debug)]
enum ResolveAction {
Resolve(GlobalValue),
/// Perform an add on the current address
Add(DataValue),
/// Load From the current address and replace it with the loaded value
Load {
/// Offset added to the base pointer before doing the load.
offset: i32,
/// Type of the loaded value.
global_type: Type,
},
}
let func = self.get_current_function();
// We start with a sentinel value that will fail if we try to load / add to it
// without resolving the base GV First.
let mut current_val = DataValue::I8(0);
let mut action_stack = vec![ResolveAction::Resolve(gv)];
loop {
match action_stack.pop() {
Some(ResolveAction::Resolve(gv)) => match func.global_values[gv] {
GlobalValueData::VMContext => {
// Fetch the VMContext value from the values of the first block in the function
let index = func
.signature
.params
.iter()
.enumerate()
.find(|(_, p)| p.purpose == ArgumentPurpose::VMContext)
.map(|(i, _)| i)
// This should be validated by the verifier
.expect("No VMCtx argument was found, but one is referenced");
let first_block =
func.layout.blocks().next().expect("to have a first block");
let vmctx_value = func.dfg.block_params(first_block)[index];
current_val = self.current_frame().get(vmctx_value).clone();
}
GlobalValueData::Load {
base,
offset,
global_type,
..
} => {
action_stack.push(ResolveAction::Load {
offset: offset.into(),
global_type,
});
action_stack.push(ResolveAction::Resolve(base));
}
GlobalValueData::IAddImm {
base,
offset,
global_type,
} => {
let offset: i64 = offset.into();
let dv = DataValue::int(offset as i128, global_type)
.map_err(|_| MemoryError::InvalidAddressType(global_type))?;
action_stack.push(ResolveAction::Add(dv));
action_stack.push(ResolveAction::Resolve(base));
}
GlobalValueData::Symbol { .. } => unimplemented!(),
GlobalValueData::DynScaleTargetConst { .. } => unimplemented!(),
},
Some(ResolveAction::Add(dv)) => {
current_val = current_val
.add(dv.clone())
.map_err(|_| MemoryError::InvalidAddress(dv))?;
}
Some(ResolveAction::Load {
offset,
global_type,
}) => {
let mut addr = Address::try_from(current_val)?;
// We can forego bounds checking here since its performed in `checked_load`
addr.offset += offset as u64;
current_val = self.checked_load(addr, global_type)?;
}
// We are done resolving this, return the current value
None => return Ok(current_val),
}
}
}
fn validate_address(&self, addr: &Address) -> Result<(), MemoryError> {
match addr.region {
AddressRegion::Stack => {
let stack_len = self.stack.len() as u64;
if addr.offset > stack_len {
return Err(MemoryError::InvalidEntry {
entry: addr.entry,
max: self.stack.len() as u64,
});
}
}
_ => unimplemented!(),
};
Ok(())
}
fn get_pinned_reg(&self) -> DataValue {
self.pinned_reg.clone()
}
fn set_pinned_reg(&mut self, v: DataValue) {
self.pinned_reg = v;
}
}
#[cfg(test)]
mod tests {
use super::*;
use crate::step::CraneliftTrap;
use cranelift_codegen::ir::immediates::Ieee32;
use cranelift_codegen::ir::TrapCode;
use cranelift_reader::parse_functions;
use smallvec::smallvec;
// Most interpreter tests should use the more ergonomic `test interpret` filetest but this
// unit test serves as a sanity check that the interpreter still works without all of the
// filetest infrastructure.
#[test]
fn sanity() {
let code = "function %test() -> i8 {
block0:
v0 = iconst.i32 1
v1 = iadd_imm v0, 1
v2 = irsub_imm v1, 44 ; 44 - 2 == 42 (see irsub_imm's semantics)
v3 = icmp_imm eq v2, 42
return v3
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let result = Interpreter::new(state)
.call_by_name("%test", &[])
.unwrap()
.unwrap_return();
assert_eq!(result, vec![DataValue::I8(1)])
}
// We don't have a way to check for traps with the current filetest infrastructure
#[test]
fn udiv_by_zero_traps() {
let code = "function %test() -> i32 {
block0:
v0 = iconst.i32 1
v1 = udiv_imm.i32 v0, 0
return v1
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let trap = Interpreter::new(state)
.call_by_name("%test", &[])
.unwrap()
.unwrap_trap();
assert_eq!(trap, CraneliftTrap::User(TrapCode::IntegerDivisionByZero));
}
#[test]
fn sdiv_min_by_neg_one_traps_with_overflow() {
let code = "function %test() -> i8 {
block0:
v0 = iconst.i32 -2147483648
v1 = sdiv_imm.i32 v0, -1
return v1
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let result = Interpreter::new(state).call_by_name("%test", &[]).unwrap();
match result {
ControlFlow::Trap(CraneliftTrap::User(TrapCode::IntegerOverflow)) => {}
_ => panic!("Unexpected ControlFlow: {:?}", result),
}
}
// This test verifies that functions can refer to each other using the function store. A double indirection is
// required, which is tricky to get right: a referenced function is a FuncRef when called but a FuncIndex inside the
// function store. This test would preferably be a CLIF filetest but the filetest infrastructure only looks at a
// single function at a time--we need more than one function in the store for this test.
#[test]
fn function_references() {
let code = "
function %child(i32) -> i32 {
block0(v0: i32):
v1 = iadd_imm v0, -1
return v1
}
function %parent(i32) -> i32 {
fn42 = %child(i32) -> i32
block0(v0: i32):
v1 = iadd_imm v0, 1
v2 = call fn42(v1)
return v2
}";
let mut env = FunctionStore::default();
let funcs = parse_functions(code).unwrap().to_vec();
funcs.iter().for_each(|f| env.add(f.name.to_string(), f));
let state = InterpreterState::default().with_function_store(env);
let result = Interpreter::new(state)
.call_by_name("%parent", &[DataValue::I32(0)])
.unwrap()
.unwrap_return();
assert_eq!(result, vec![DataValue::I32(0)])
}
#[test]
fn fuel() {
let code = "function %test() -> i8 {
block0:
v0 = iconst.i32 1
v1 = iadd_imm v0, 1
return v1
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
// The default interpreter should not enable the fuel mechanism
let state = InterpreterState::default().with_function_store(env.clone());
let result = Interpreter::new(state)
.call_by_name("%test", &[])
.unwrap()
.unwrap_return();
assert_eq!(result, vec![DataValue::I32(2)]);
// With 2 fuel, we should execute the iconst and iadd, but not the return thus giving a
// fuel exhausted error
let state = InterpreterState::default().with_function_store(env.clone());
let result = Interpreter::new(state)
.with_fuel(Some(2))
.call_by_name("%test", &[]);
match result {
Err(InterpreterError::FuelExhausted) => {}
_ => panic!("Expected Err(FuelExhausted), but got {:?}", result),
}
// With 3 fuel, we should be able to execute the return instruction, and complete the test
let state = InterpreterState::default().with_function_store(env.clone());
let result = Interpreter::new(state)
.with_fuel(Some(3))
.call_by_name("%test", &[])
.unwrap()
.unwrap_return();
assert_eq!(result, vec![DataValue::I32(2)]);
}
// Verifies that writing to the stack on a called function does not overwrite the parents
// stack slots.
#[test]
fn stack_slots_multi_functions() {
let code = "
function %callee(i64, i64) -> i64 {
ss0 = explicit_slot 8
ss1 = explicit_slot 8
block0(v0: i64, v1: i64):
stack_store.i64 v0, ss0
stack_store.i64 v1, ss1
v2 = stack_load.i64 ss0
v3 = stack_load.i64 ss1
v4 = iadd.i64 v2, v3
return v4
}
function %caller(i64, i64, i64, i64) -> i64 {
fn0 = %callee(i64, i64) -> i64
ss0 = explicit_slot 8
ss1 = explicit_slot 8
block0(v0: i64, v1: i64, v2: i64, v3: i64):
stack_store.i64 v0, ss0
stack_store.i64 v1, ss1
v4 = call fn0(v2, v3)
v5 = stack_load.i64 ss0
v6 = stack_load.i64 ss1
v7 = iadd.i64 v4, v5
v8 = iadd.i64 v7, v6
return v8
}";
let mut env = FunctionStore::default();
let funcs = parse_functions(code).unwrap().to_vec();
funcs.iter().for_each(|f| env.add(f.name.to_string(), f));
let state = InterpreterState::default().with_function_store(env);
let result = Interpreter::new(state)
.call_by_name(
"%caller",
&[
DataValue::I64(3),
DataValue::I64(5),
DataValue::I64(7),
DataValue::I64(11),
],
)
.unwrap()
.unwrap_return();
assert_eq!(result, vec![DataValue::I64(26)])
}
#[test]
fn out_of_slot_write_traps() {
let code = "
function %stack_write() {
ss0 = explicit_slot 8
block0:
v0 = iconst.i64 10
stack_store.i64 v0, ss0+8
return
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let trap = Interpreter::new(state)
.call_by_name("%stack_write", &[])
.unwrap()
.unwrap_trap();
assert_eq!(trap, CraneliftTrap::User(TrapCode::HeapOutOfBounds));
}
#[test]
fn partial_out_of_slot_write_traps() {
let code = "
function %stack_write() {
ss0 = explicit_slot 8
block0:
v0 = iconst.i64 10
stack_store.i64 v0, ss0+4
return
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let trap = Interpreter::new(state)
.call_by_name("%stack_write", &[])
.unwrap()
.unwrap_trap();
assert_eq!(trap, CraneliftTrap::User(TrapCode::HeapOutOfBounds));
}
#[test]
fn out_of_slot_read_traps() {
let code = "
function %stack_load() {
ss0 = explicit_slot 8
block0:
v0 = stack_load.i64 ss0+8
return
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let trap = Interpreter::new(state)
.call_by_name("%stack_load", &[])
.unwrap()
.unwrap_trap();
assert_eq!(trap, CraneliftTrap::User(TrapCode::HeapOutOfBounds));
}
#[test]
fn partial_out_of_slot_read_traps() {
let code = "
function %stack_load() {
ss0 = explicit_slot 8
block0:
v0 = stack_load.i64 ss0+4
return
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let trap = Interpreter::new(state)
.call_by_name("%stack_load", &[])
.unwrap()
.unwrap_trap();
assert_eq!(trap, CraneliftTrap::User(TrapCode::HeapOutOfBounds));
}
#[test]
fn partial_out_of_slot_read_by_addr_traps() {
let code = "
function %stack_load() {
ss0 = explicit_slot 8
block0:
v0 = stack_addr.i64 ss0
v1 = iconst.i64 4
v2 = iadd.i64 v0, v1
v3 = load.i64 v2
return
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let trap = Interpreter::new(state)
.call_by_name("%stack_load", &[])
.unwrap()
.unwrap_trap();
assert_eq!(trap, CraneliftTrap::User(TrapCode::HeapOutOfBounds));
}
#[test]
fn partial_out_of_slot_write_by_addr_traps() {
let code = "
function %stack_store() {
ss0 = explicit_slot 8
block0:
v0 = stack_addr.i64 ss0
v1 = iconst.i64 4
v2 = iadd.i64 v0, v1
store.i64 v1, v2
return
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let trap = Interpreter::new(state)
.call_by_name("%stack_store", &[])
.unwrap()
.unwrap_trap();
assert_eq!(trap, CraneliftTrap::User(TrapCode::HeapOutOfBounds));
}
#[test]
fn srem_trap() {
let code = "function %test() -> i64 {
block0:
v0 = iconst.i64 0x8000_0000_0000_0000
v1 = iconst.i64 -1
v2 = srem.i64 v0, v1
return v2
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default().with_function_store(env);
let trap = Interpreter::new(state)
.call_by_name("%test", &[])
.unwrap()
.unwrap_trap();
assert_eq!(trap, CraneliftTrap::User(TrapCode::IntegerOverflow));
}
#[test]
fn libcall() {
let code = "function %test() -> i64 {
fn0 = colocated %CeilF32 (f32) -> f32 fast
block0:
v1 = f32const 0x0.5
v2 = call fn0(v1)
return v2
}";
let func = parse_functions(code).unwrap().into_iter().next().unwrap();
let mut env = FunctionStore::default();
env.add(func.name.to_string(), &func);
let state = InterpreterState::default()
.with_function_store(env)
.with_libcall_handler(|libcall, args| {
Ok(smallvec![match (libcall, &args[..]) {
(LibCall::CeilF32, [DataValue::F32(a)]) => DataValue::F32(a.ceil()),
_ => panic!("Unexpected args"),
}])
});
let result = Interpreter::new(state)
.call_by_name("%test", &[])
.unwrap()
.unwrap_return();
assert_eq!(result, vec![DataValue::F32(Ieee32::with_float(1.0))])
}
}