* Refactor where results of compilation are stored This commit refactors the internals of compilation in Wasmtime to change where results of individual function compilation are stored. Previously compilation resulted in many maps being returned, and compilation results generally held all these maps together. This commit instead switches this to have all metadata stored in a `CompiledFunction` instead of having a separate map for each item that can be stored. The motivation for this is primarily to help out with future module-linking-related PRs. What exactly "module level" is depends on how we interpret modules and how many modules are in play, so it's a bit easier for operations in wasmtime to work at the function level where possible. This means that we don't have to pass around multiple different maps and a function index, but instead just one map or just one entry representing a compiled function. Additionally this change updates where the parallelism of compilation happens, pushing it into `wasmtime-jit` instead of `wasmtime-environ`. This is another goal where `wasmtime-jit` will have more knowledge about module-level pieces with module linking in play. User-facing-wise this should be the same in terms of parallel compilation, though. The ultimate goal of this refactoring is to make it easier for the results of compilation to actually be a set of wasm modules. This means we won't be able to have a map-per-metadata where the primary key is the function index, because there will be many modules within one "object file". * Don't clear out fields, just don't store them Persist a smaller set of fields in `CompilationArtifacts` instead of trying to clear fields out and dynamically not accessing them.
184 lines
5.6 KiB
Rust
184 lines
5.6 KiB
Rust
//! JIT compilation.
|
|
|
|
use crate::instantiate::SetupError;
|
|
use crate::object::{build_object, ObjectUnwindInfo};
|
|
use object::write::Object;
|
|
use std::hash::{Hash, Hasher};
|
|
use wasmtime_debug::{emit_dwarf, DwarfSection};
|
|
use wasmtime_environ::entity::EntityRef;
|
|
use wasmtime_environ::isa::{TargetFrontendConfig, TargetIsa};
|
|
use wasmtime_environ::wasm::{DefinedMemoryIndex, MemoryIndex};
|
|
use wasmtime_environ::{
|
|
CompiledFunctions, Compiler as EnvCompiler, DebugInfoData, Module, ModuleMemoryOffset,
|
|
ModuleTranslation, Tunables, VMOffsets,
|
|
};
|
|
|
|
/// Select which kind of compilation to use.
|
|
#[derive(Copy, Clone, Debug, Hash)]
|
|
pub enum CompilationStrategy {
|
|
/// Let Wasmtime pick the strategy.
|
|
Auto,
|
|
|
|
/// Compile all functions with Cranelift.
|
|
Cranelift,
|
|
|
|
/// Compile all functions with Lightbeam.
|
|
#[cfg(feature = "lightbeam")]
|
|
Lightbeam,
|
|
}
|
|
|
|
/// A WebAssembly code JIT compiler.
|
|
///
|
|
/// A `Compiler` instance owns the executable memory that it allocates.
|
|
///
|
|
/// TODO: Evolve this to support streaming rather than requiring a `&[u8]`
|
|
/// containing a whole wasm module at once.
|
|
///
|
|
/// TODO: Consider using cranelift-module.
|
|
pub struct Compiler {
|
|
isa: Box<dyn TargetIsa>,
|
|
compiler: Box<dyn EnvCompiler>,
|
|
strategy: CompilationStrategy,
|
|
tunables: Tunables,
|
|
}
|
|
|
|
impl Compiler {
|
|
/// Construct a new `Compiler`.
|
|
pub fn new(isa: Box<dyn TargetIsa>, strategy: CompilationStrategy, tunables: Tunables) -> Self {
|
|
Self {
|
|
isa,
|
|
strategy,
|
|
compiler: match strategy {
|
|
CompilationStrategy::Auto | CompilationStrategy::Cranelift => {
|
|
Box::new(wasmtime_environ::cranelift::Cranelift::default())
|
|
}
|
|
#[cfg(feature = "lightbeam")]
|
|
CompilationStrategy::Lightbeam => Box::new(wasmtime_environ::lightbeam::Lightbeam),
|
|
},
|
|
tunables,
|
|
}
|
|
}
|
|
}
|
|
|
|
fn _assert_compiler_send_sync() {
|
|
fn _assert<T: Send + Sync>() {}
|
|
_assert::<Compiler>();
|
|
}
|
|
|
|
fn transform_dwarf_data(
|
|
isa: &dyn TargetIsa,
|
|
module: &Module,
|
|
debug_data: &DebugInfoData,
|
|
funcs: &CompiledFunctions,
|
|
) -> Result<Vec<DwarfSection>, SetupError> {
|
|
let target_config = isa.frontend_config();
|
|
let ofs = VMOffsets::new(target_config.pointer_bytes(), &module.local);
|
|
|
|
let memory_offset = if ofs.num_imported_memories > 0 {
|
|
ModuleMemoryOffset::Imported(ofs.vmctx_vmmemory_import(MemoryIndex::new(0)))
|
|
} else if ofs.num_defined_memories > 0 {
|
|
ModuleMemoryOffset::Defined(ofs.vmctx_vmmemory_definition_base(DefinedMemoryIndex::new(0)))
|
|
} else {
|
|
ModuleMemoryOffset::None
|
|
};
|
|
emit_dwarf(isa, debug_data, funcs, &memory_offset).map_err(SetupError::DebugInfo)
|
|
}
|
|
|
|
#[allow(missing_docs)]
|
|
pub struct Compilation {
|
|
pub obj: Object,
|
|
pub unwind_info: Vec<ObjectUnwindInfo>,
|
|
pub funcs: CompiledFunctions,
|
|
}
|
|
|
|
impl Compiler {
|
|
/// Return the isa.
|
|
pub fn isa(&self) -> &dyn TargetIsa {
|
|
self.isa.as_ref()
|
|
}
|
|
|
|
/// Return the target's frontend configuration settings.
|
|
pub fn frontend_config(&self) -> TargetFrontendConfig {
|
|
self.isa.frontend_config()
|
|
}
|
|
|
|
/// Return the tunables in use by this engine.
|
|
pub fn tunables(&self) -> &Tunables {
|
|
&self.tunables
|
|
}
|
|
|
|
/// Compile the given function bodies.
|
|
pub fn compile<'data>(
|
|
&self,
|
|
translation: &ModuleTranslation,
|
|
) -> Result<Compilation, SetupError> {
|
|
cfg_if::cfg_if! {
|
|
if #[cfg(feature = "parallel-compilation")] {
|
|
use rayon::prelude::*;
|
|
let iter = translation.function_body_inputs
|
|
.iter()
|
|
.collect::<Vec<_>>()
|
|
.into_par_iter();
|
|
} else {
|
|
let iter = translation.function_body_inputs.iter();
|
|
}
|
|
}
|
|
let funcs = iter
|
|
.map(|(index, func)| {
|
|
self.compiler
|
|
.compile_function(translation, index, func, &*self.isa)
|
|
})
|
|
.collect::<Result<Vec<_>, _>>()?
|
|
.into_iter()
|
|
.collect::<CompiledFunctions>();
|
|
|
|
let dwarf_sections = if translation.debuginfo.is_some() && !funcs.is_empty() {
|
|
transform_dwarf_data(
|
|
&*self.isa,
|
|
&translation.module,
|
|
translation.debuginfo.as_ref().unwrap(),
|
|
&funcs,
|
|
)?
|
|
} else {
|
|
vec![]
|
|
};
|
|
|
|
let (obj, unwind_info) =
|
|
build_object(&*self.isa, &translation.module, &funcs, dwarf_sections)?;
|
|
|
|
Ok(Compilation {
|
|
obj,
|
|
unwind_info,
|
|
funcs,
|
|
})
|
|
}
|
|
}
|
|
|
|
impl Hash for Compiler {
|
|
fn hash<H: Hasher>(&self, hasher: &mut H) {
|
|
let Compiler {
|
|
strategy,
|
|
compiler: _,
|
|
isa,
|
|
tunables,
|
|
} = self;
|
|
|
|
// Hash compiler's flags: compilation strategy, isa, frontend config,
|
|
// misc tunables.
|
|
strategy.hash(hasher);
|
|
isa.triple().hash(hasher);
|
|
// TODO: if this `to_string()` is too expensive then we should upstream
|
|
// a native hashing ability of flags into cranelift itself, but
|
|
// compilation and/or cache loading is relatively expensive so seems
|
|
// unlikely.
|
|
isa.flags().to_string().hash(hasher);
|
|
isa.frontend_config().hash(hasher);
|
|
tunables.hash(hasher);
|
|
|
|
// TODO: ... and should we hash anything else? There's a lot of stuff in
|
|
// `TargetIsa`, like registers/encodings/etc. Should we be hashing that
|
|
// too? It seems like wasmtime doesn't configure it too too much, but
|
|
// this may become an issue at some point.
|
|
}
|
|
}
|