Have vmctx be the first argument so we (almost) never have to shuffle it around

This commit is contained in:
Jef
2019-01-16 09:33:46 +01:00
parent b86d11e4ad
commit d7434fe5d2
4 changed files with 34 additions and 21 deletions

View File

@@ -428,7 +428,7 @@ impl Locals {
}
fn vmctx_index(&self) -> u32 {
self.num_args() - 1
0
}
}
@@ -1226,7 +1226,11 @@ impl Context<'_> {
self.block_state.return_register = return_reg;
}
pub fn end_block(&mut self, parent_block_state: BlockState, func: impl FnOnce(&mut Self)) {
pub fn end_block(
&mut self,
parent_block_state: BlockState,
before_push_return: impl FnOnce(&mut Self),
) {
// TODO: This should currently never be called, but is important for if we want to
// have a more complex stack spilling scheme.
debug_assert_eq!(
@@ -1246,7 +1250,7 @@ impl Context<'_> {
self.block_state = parent_block_state;
self.block_state.locals = locals;
func(self);
before_push_return(self);
if let Some(reg) = return_reg {
self.block_state.regs.mark_used(reg);
@@ -1800,15 +1804,11 @@ impl Context<'_> {
/// Unfortunately, we can't elide this store if we're just passing arguments on
/// because these registers are caller-saved and so the callee can use them as
/// scratch space.
fn free_arg_registers(&mut self, arity: u32) {
fn free_arg_registers(&mut self, exclude: Option<u32>) {
// This is bound to the maximum size of the `ArrayVec` amd so can be considered to have constant
// runtime
for i in 0..self
.block_state
.locals
.register_arguments
.len()
.min(arity as usize)
for i in (0..self.block_state.locals.register_arguments.len())
.filter(|i| exclude != Some(*i as u32))
{
match self.block_state.locals.register_arguments[i] {
ArgLoc::Register(reg) => {
@@ -1916,7 +1916,7 @@ impl Context<'_> {
) -> CallCleanup {
let num_stack_args = (arity as usize).saturating_sub(ARGS_IN_GPRS.len()) as i32;
self.free_arg_registers(if has_vmctx { arity - 1} else { arity } );
self.free_arg_registers(if has_vmctx { Some(0) } else { None });
// We pop stack arguments first - arguments are RTL
if num_stack_args > 0 {
@@ -2002,8 +2002,14 @@ impl Context<'_> {
"We don't support multiple return yet"
);
let vmctx = Value::Local(self.block_state.locals.vmctx_index());
self.push(vmctx);
let vmctx = StackValue::Local(self.block_state.locals.vmctx_index());
let count = self.block_state.stack.len();
// TODO: I believe that this can't cause quadratic runtime but I'm not
// certain.
self.block_state
.stack
.insert(count - arg_arity as usize, vmctx);
let cleanup = self.pass_outgoing_args(arg_arity + 1, return_arity, true);
let label = &self.func_starts[index as usize].1;
@@ -2020,6 +2026,8 @@ impl Context<'_> {
// TODO: Allow use of unused argument registers as scratch registers.
/// Writes the function prologue and stores the arguments as locals
pub fn start_function(&mut self, arguments: u32, locals: u32) -> FunctionEnd {
// To support `vmctx`
let arguments = arguments + 1;
let reg_args = &ARGS_IN_GPRS[..(arguments as usize).min(ARGS_IN_GPRS.len())];
// We need space to store the register arguments if we need to call a function

View File

@@ -158,7 +158,7 @@ pub fn translate(
let operators = body.get_operators_reader()?;
// We must add 1 here to supply `vmctx`
let func = ctx.start_function(arg_count + 1, num_locals);
let func = ctx.start_function(arg_count, num_locals);
let mut control_frames = Vec::new();

View File

@@ -62,11 +62,9 @@ macro_rules! impl_function_args {
impl<$first, $($rest),*> FunctionArgs for ($first, $($rest),*) {
#[allow(non_snake_case)]
unsafe fn call<T>(self, start: *const u8, vm_ctx: *const u8) -> T {
let func = mem::transmute::<_, extern "sysv64" fn($first $(, $rest)*, VmCtx) -> T>(start);
{
let ($first, $($rest),*) = self;
func($first $(, $rest)*, vm_ctx as VmCtx)
}
let func = mem::transmute::<_, extern "sysv64" fn(VmCtx, $first $(, $rest)*) -> T>(start);
let ($first, $($rest),*) = self;
func(vm_ctx as VmCtx, $first $(, $rest)*)
}
}

View File

@@ -10,6 +10,7 @@ fn translate_wat(wat: &str) -> TranslatedModule {
/// Execute the first function in the module.
fn execute_wat(wat: &str, a: u32, b: u32) -> u32 {
let translated = translate_wat(wat);
translated.disassemble();
translated.execute_func(0, (a, b)).unwrap()
}
@@ -317,6 +318,7 @@ fn function_read_args_spill_to_stack() {
let code = r#"
(module
(func (param i32) (param i32) (param i32) (param i32)
(param i32) (param i32) (param i32) (param i32)
(param i32) (param i32) (param i32) (param i32)
(result i32)
@@ -339,7 +341,12 @@ fn function_read_args_spill_to_stack() {
{
let translated = translate_wat(code);
translated.disassemble();
translated.execute_func(0, (7u32, 6u32, 5u32, 4u32, 3u32, 2u32, 1u32, 0u32))
translated.execute_func(
0,
(
7u32, 6u32, 5u32, 4u32, 3u32, 2u32, 1u32, 0u32, 1u32, 2u32, 3u32, 4u32,
),
)
},
Ok(7u32)
);
@@ -414,6 +421,7 @@ macro_rules! mk_function_write_args_spill_to_stack {
}
};
}
mk_function_write_args_spill_to_stack!(function_write_args_spill_to_stack_i32, i32);
mk_function_write_args_spill_to_stack!(function_write_args_spill_to_stack_i64, i64);
@@ -539,7 +547,6 @@ fn spec_loop() {
translated.execute_func::<(), ()>(0, ()).unwrap();
}
quickcheck! {
fn spec_fac(n: i8) -> bool {
const CODE: &str = r#"