Compare commits

..

3 Commits

Author SHA1 Message Date
T0b1
3f93c5f698 changes 2023-04-15 05:04:07 +02:00
T0b1
afb8d61c87 changes 2023-04-15 05:00:52 +02:00
T0b1
a31407c991 maybe better blockorder impl 2023-04-15 04:18:14 +02:00
10 changed files with 65 additions and 70 deletions

View File

@@ -84,9 +84,6 @@ anyhow = { workspace = true }
[profile.release.build-override] [profile.release.build-override]
opt-level = 0 opt-level = 0
[profile.release]
debug = true
[workspace] [workspace]
resolver = '2' resolver = '2'
members = [ members = [

View File

@@ -27,7 +27,7 @@ serde = { version = "1.0.94", features = ["derive"], optional = true }
bincode = { version = "1.2.1", optional = true } bincode = { version = "1.2.1", optional = true }
gimli = { workspace = true, features = ["write"], optional = true } gimli = { workspace = true, features = ["write"], optional = true }
smallvec = { workspace = true } smallvec = { workspace = true }
regalloc2 = { path = "../../../regalloc2/", features = ["checker", "trace-log"] } regalloc2 = { version = "0.6.1", features = ["checker"] }
souper-ir = { version = "2.1.0", optional = true } souper-ir = { version = "2.1.0", optional = true }
sha2 = { version = "0.10.2", optional = true } sha2 = { version = "0.10.2", optional = true }
# It is a goal of the cranelift-codegen crate to have minimal external dependencies. # It is a goal of the cranelift-codegen crate to have minimal external dependencies.

View File

@@ -230,7 +230,6 @@ pub fn create_reg_env(flags: &settings::Flags) -> MachineEnv {
], ],
], ],
fixed_stack_slots: vec![], fixed_stack_slots: vec![],
scratch_by_class: [None, None],
}; };
if !flags.enable_pinned_reg() { if !flags.enable_pinned_reg() {

View File

@@ -182,7 +182,6 @@ pub fn crate_reg_eviroment(_flags: &settings::Flags) -> MachineEnv {
preferred_regs_by_class, preferred_regs_by_class,
non_preferred_regs_by_class, non_preferred_regs_by_class,
fixed_stack_slots: vec![], fixed_stack_slots: vec![],
scratch_by_class: [None, None],
} }
} }

View File

@@ -151,7 +151,6 @@ pub fn create_machine_env(_flags: &settings::Flags) -> MachineEnv {
], ],
], ],
fixed_stack_slots: vec![], fixed_stack_slots: vec![],
scratch_by_class: [None, None],
} }
} }

View File

@@ -203,7 +203,6 @@ pub(crate) fn create_reg_env_systemv(flags: &settings::Flags) -> MachineEnv {
vec![], vec![],
], ],
fixed_stack_slots: vec![], fixed_stack_slots: vec![],
scratch_by_class: [None, None]
}; };
debug_assert_eq!(r15(), pinned_reg()); debug_assert_eq!(r15(), pinned_reg());

View File

@@ -81,15 +81,6 @@ pub struct BlockLoweringOrder {
/// Ranges in `lowered_succ_indices` giving the successor lists for each lowered /// Ranges in `lowered_succ_indices` giving the successor lists for each lowered
/// block. Indexed by lowering-order index (`BlockIndex`). /// block. Indexed by lowering-order index (`BlockIndex`).
lowered_succ_ranges: Vec<(Option<Inst>, std::ops::Range<usize>)>, lowered_succ_ranges: Vec<(Option<Inst>, std::ops::Range<usize>)>,
/// Cold blocks. These blocks are not reordered in the
/// `lowered_order` above; the lowered order must respect RPO
/// (uses after defs) in order for lowering to be
/// correct. Instead, this set is used to provide `is_cold()`,
/// which is used by VCode emission to sink the blocks at the last
/// moment (when we actually emit bytes into the MachBuffer).
cold_blocks: FxHashSet<BlockIndex>,
/// Lowered blocks that are indirect branch targets.
indirect_branch_targets: FxHashSet<BlockIndex>,
} }
#[derive(Clone, Copy, Debug, PartialEq, Eq, Hash)] #[derive(Clone, Copy, Debug, PartialEq, Eq, Hash)]
@@ -98,6 +89,15 @@ pub enum LoweredBlock {
Orig { Orig {
/// Original CLIF block. /// Original CLIF block.
block: Block, block: Block,
/// Cold block. These blocks are not reordered in the
/// `lowered_order` above; the lowered order must respect RPO
/// (uses after defs) in order for lowering to be
/// correct. Instead, this bool is used to provide `is_cold()`,
/// which is used by VCode emission to sink the blocks at the last
/// moment (when we actually emit bytes into the MachBuffer).
cold: bool,
/// Target of indirect branch
indirect_branch_target: bool,
}, },
/// Critical edge between two CLIF blocks. /// Critical edge between two CLIF blocks.
@@ -112,6 +112,11 @@ pub enum LoweredBlock {
/// indexing order as `inst_predicates::visit_block_succs`. This is used to distinguish /// indexing order as `inst_predicates::visit_block_succs`. This is used to distinguish
/// multiple edges between the same CLIF blocks. /// multiple edges between the same CLIF blocks.
succ_idx: u32, succ_idx: u32,
/// See [LoweredBlock::orig]
cold: bool,
/// See [LoweredBlock::orig]
indirect_branch_target: bool,
}, },
} }
@@ -119,7 +124,7 @@ impl LoweredBlock {
/// Unwrap an `Orig` block. /// Unwrap an `Orig` block.
pub fn orig_block(&self) -> Option<Block> { pub fn orig_block(&self) -> Option<Block> {
match self { match self {
&LoweredBlock::Orig { block } => Some(block), &LoweredBlock::Orig { block, .. } => Some(block),
&LoweredBlock::CriticalEdge { .. } => None, &LoweredBlock::CriticalEdge { .. } => None,
} }
} }
@@ -141,6 +146,20 @@ impl LoweredBlock {
&LoweredBlock::Orig { .. } => None, &LoweredBlock::Orig { .. } => None,
} }
} }
pub fn is_cold(&self) -> bool {
match self {
&LoweredBlock::CriticalEdge { cold, .. } => cold,
&LoweredBlock::Orig { cold, .. } => cold
}
}
pub fn is_indirect_branch_target(&self) -> bool {
match self {
&LoweredBlock::CriticalEdge { indirect_branch_target, .. } => indirect_branch_target,
&LoweredBlock::Orig { indirect_branch_target, .. } => indirect_branch_target
}
}
} }
impl BlockLoweringOrder { impl BlockLoweringOrder {
@@ -165,7 +184,7 @@ impl BlockLoweringOrder {
visit_block_succs(f, block, |_, succ, from_table| { visit_block_succs(f, block, |_, succ, from_table| {
block_out_count[block] += 1; block_out_count[block] += 1;
block_in_count[succ] += 1; block_in_count[succ] += 1;
block_succs.push(LoweredBlock::Orig { block: succ }); block_succs.push(LoweredBlock::Orig { block: succ, cold: f.layout.is_cold(succ), indirect_branch_target: false });
if from_table { if from_table {
indirect_branch_target_clif_blocks.insert(succ); indirect_branch_target_clif_blocks.insert(succ);
@@ -192,7 +211,7 @@ impl BlockLoweringOrder {
let mut lowered_order = Vec::new(); let mut lowered_order = Vec::new();
for &block in domtree.cfg_postorder().iter().rev() { for &block in domtree.cfg_postorder().iter().rev() {
let lb = LoweredBlock::Orig { block }; let lb = LoweredBlock::Orig { block, cold: f.layout.is_cold(block), indirect_branch_target: indirect_branch_target_clif_blocks.contains(&block) };
let bindex = BlockIndex::new(lowered_order.len()); let bindex = BlockIndex::new(lowered_order.len());
lb_to_bindex.insert(lb.clone(), bindex); lb_to_bindex.insert(lb.clone(), bindex);
lowered_order.push(lb); lowered_order.push(lb);
@@ -208,6 +227,10 @@ impl BlockLoweringOrder {
pred: block, pred: block,
succ, succ,
succ_idx: succ_ix as u32, succ_idx: succ_ix as u32,
// Edges inherit indirect branch and cold block metadata from their
// successor.
cold: f.layout.is_cold(succ),
indirect_branch_target: indirect_branch_target_clif_blocks.contains(&succ)
}; };
let bindex = BlockIndex::new(lowered_order.len()); let bindex = BlockIndex::new(lowered_order.len());
lb_to_bindex.insert(*lb, bindex); lb_to_bindex.insert(*lb, bindex);
@@ -221,8 +244,6 @@ impl BlockLoweringOrder {
// during the creation of `lowering_order`, as we need `lb_to_bindex` to be fully populated // during the creation of `lowering_order`, as we need `lb_to_bindex` to be fully populated
// first. // first.
let mut lowered_succ_indices = Vec::new(); let mut lowered_succ_indices = Vec::new();
let mut cold_blocks = FxHashSet::default();
let mut indirect_branch_targets = FxHashSet::default();
let lowered_succ_ranges = let lowered_succ_ranges =
Vec::from_iter(lowered_order.iter().enumerate().map(|(ix, lb)| { Vec::from_iter(lowered_order.iter().enumerate().map(|(ix, lb)| {
let bindex = BlockIndex::new(ix); let bindex = BlockIndex::new(ix);
@@ -230,18 +251,24 @@ impl BlockLoweringOrder {
let opt_inst = match lb { let opt_inst = match lb {
// Block successors are pulled directly over, as they'll have been mutated when // Block successors are pulled directly over, as they'll have been mutated when
// determining the block order already. // determining the block order already.
&LoweredBlock::Orig { block } => { &LoweredBlock::Orig { block, .. } => {
let range = block_succ_range[block].clone(); let range = block_succ_range[block].clone();
lowered_succ_indices lowered_succ_indices
.extend(block_succs[range].iter().map(|lb| lb_to_bindex[lb])); .extend(block_succs[range].iter().map(|lb| {
match lb {
if f.layout.is_cold(block) { LoweredBlock::Orig { block, cold, .. } => {
cold_blocks.insert(bindex); match lb_to_bindex.get(&LoweredBlock::Orig { block: *block, cold: *cold, indirect_branch_target: indirect_branch_target_clif_blocks.contains(block) } ) {
Some(idx) => *idx,
None => {
panic!("Trying to hashlookup {:?} in HashMap: {:?}", lb, lb_to_bindex);
} }
if indirect_branch_target_clif_blocks.contains(&block) {
indirect_branch_targets.insert(bindex);
} }
},
LoweredBlock::CriticalEdge { .. } => {
lb_to_bindex[&lb]
}
}
}));
let last = f.layout.last_inst(block).unwrap(); let last = f.layout.last_inst(block).unwrap();
let opcode = f.dfg.insts[last].opcode(); let opcode = f.dfg.insts[last].opcode();
@@ -253,21 +280,10 @@ impl BlockLoweringOrder {
// Critical edges won't have successor information in block_succ_range, but // Critical edges won't have successor information in block_succ_range, but
// they only have a single known successor to record anyway. // they only have a single known successor to record anyway.
&LoweredBlock::CriticalEdge { succ, .. } => { &LoweredBlock::CriticalEdge { succ, cold, indirect_branch_target, .. } => {
let succ_index = lb_to_bindex[&LoweredBlock::Orig { block: succ }]; let succ_index = lb_to_bindex[&LoweredBlock::Orig { block: succ, cold, indirect_branch_target }];
lowered_succ_indices.push(succ_index); lowered_succ_indices.push(succ_index);
// Edges inherit indirect branch and cold block metadata from their
// successor.
if f.layout.is_cold(succ) {
cold_blocks.insert(bindex);
}
if indirect_branch_target_clif_blocks.contains(&succ) {
indirect_branch_targets.insert(bindex);
}
None None
} }
}; };
@@ -278,9 +294,7 @@ impl BlockLoweringOrder {
let result = BlockLoweringOrder { let result = BlockLoweringOrder {
lowered_order, lowered_order,
lowered_succ_indices, lowered_succ_indices,
lowered_succ_ranges, lowered_succ_ranges
cold_blocks,
indirect_branch_targets,
}; };
trace!("BlockLoweringOrder: {:#?}", result); trace!("BlockLoweringOrder: {:#?}", result);
@@ -300,13 +314,13 @@ impl BlockLoweringOrder {
/// Determine whether the given lowered-block index is cold. /// Determine whether the given lowered-block index is cold.
pub fn is_cold(&self, block: BlockIndex) -> bool { pub fn is_cold(&self, block: BlockIndex) -> bool {
self.cold_blocks.contains(&block) self.lowered_order[block.index()].is_cold()
} }
/// Determine whether the given lowered block index is an indirect branch /// Determine whether the given lowered block index is an indirect branch
/// target. /// target.
pub fn is_indirect_branch_target(&self, block: BlockIndex) -> bool { pub fn is_indirect_branch_target(&self, block: BlockIndex) -> bool {
self.indirect_branch_targets.contains(&block) self.lowered_order[block.index()].is_indirect_branch_target()
} }
} }

View File

@@ -51,7 +51,6 @@ pub fn compile<B: LowerBackend + TargetIsa>(
let _tt = timing::regalloc(); let _tt = timing::regalloc();
let mut options = RegallocOptions::default(); let mut options = RegallocOptions::default();
options.verbose_log = b.flags().regalloc_verbose_logs(); options.verbose_log = b.flags().regalloc_verbose_logs();
options.fast_alloc = true;
if cfg!(debug_assertions) { if cfg!(debug_assertions) {
options.validate_ssa = true; options.validate_ssa = true;

View File

@@ -908,7 +908,6 @@ impl<I: VCodeInst> VCode<I> {
for inst_or_edit in regalloc.block_insts_and_edits(&self, block) { for inst_or_edit in regalloc.block_insts_and_edits(&self, block) {
match inst_or_edit { match inst_or_edit {
InstOrEdit::Inst(iix) => { InstOrEdit::Inst(iix) => {
trace!("Emitting inst {}", iix.index());
if !self.debug_value_labels.is_empty() { if !self.debug_value_labels.is_empty() {
// If we need to produce debug info, // If we need to produce debug info,
// record the offset of each instruction // record the offset of each instruction
@@ -1010,7 +1009,6 @@ impl<I: VCodeInst> VCode<I> {
// immediately emit it. // immediately emit it.
match (from.as_reg(), to.as_reg()) { match (from.as_reg(), to.as_reg()) {
(Some(from), Some(to)) => { (Some(from), Some(to)) => {
trace!("Emitting move from {} to {}", from, to);
// Reg-to-reg move. // Reg-to-reg move.
let from_rreg = Reg::from(from); let from_rreg = Reg::from(from);
let to_rreg = Writable::from_reg(Reg::from(to)); let to_rreg = Writable::from_reg(Reg::from(to));
@@ -1020,7 +1018,6 @@ impl<I: VCodeInst> VCode<I> {
do_emit(&mv, &[], &mut disasm, &mut buffer, &mut state); do_emit(&mv, &[], &mut disasm, &mut buffer, &mut state);
} }
(Some(from), None) => { (Some(from), None) => {
trace!("Emitting move from {} to slot {}", from, to.as_stack().unwrap().index());
// Spill from register to spillslot. // Spill from register to spillslot.
let to = to.as_stack().unwrap(); let to = to.as_stack().unwrap();
let from_rreg = RealReg::from(from); let from_rreg = RealReg::from(from);
@@ -1028,7 +1025,6 @@ impl<I: VCodeInst> VCode<I> {
do_emit(&spill, &[], &mut disasm, &mut buffer, &mut state); do_emit(&spill, &[], &mut disasm, &mut buffer, &mut state);
} }
(None, Some(to)) => { (None, Some(to)) => {
trace!("Emitting move from slot {} to {}", from.as_stack().unwrap().index(), to);
// Load from spillslot to register. // Load from spillslot to register.
let from = from.as_stack().unwrap(); let from = from.as_stack().unwrap();
let to_rreg = Writable::from_reg(RealReg::from(to)); let to_rreg = Writable::from_reg(RealReg::from(to));
@@ -1104,10 +1100,6 @@ impl<I: VCodeInst> VCode<I> {
self.compute_value_labels_ranges(regalloc, &inst_offsets[..], func_body_len); self.compute_value_labels_ranges(regalloc, &inst_offsets[..], func_body_len);
let frame_size = self.abi.frame_size(); let frame_size = self.abi.frame_size();
if want_disasm {
trace!("Disassembly:\n{}", disasm);
}
EmitResult { EmitResult {
buffer, buffer,
bb_offsets, bb_offsets,
@@ -1290,6 +1282,14 @@ impl<I: VCodeInst> RegallocFunction for VCode<I> {
self.insts[insn.index()].is_safepoint() self.insts[insn.index()].is_safepoint()
} }
fn is_move(&self, insn: InsnIndex) -> Option<(Operand, Operand)> {
let (a, b) = self.is_move.get(&insn)?;
Some((
self.assert_operand_not_vreg_alias(*a),
self.assert_operand_not_vreg_alias(*b),
))
}
fn inst_operands(&self, insn: InsnIndex) -> &[Operand] { fn inst_operands(&self, insn: InsnIndex) -> &[Operand] {
let (start, end) = self.operand_ranges[insn.index()]; let (start, end) = self.operand_ranges[insn.index()];
let ret = &self.operands[start as usize..end as usize]; let ret = &self.operands[start as usize..end as usize];

View File

@@ -17,7 +17,6 @@ use std::cmp::max;
use std::collections::hash_map::Entry; use std::collections::hash_map::Entry;
use std::collections::HashMap; use std::collections::HashMap;
use thiserror::Error; use thiserror::Error;
use log::trace;
const TESTFILE_NAMESPACE: u32 = 0; const TESTFILE_NAMESPACE: u32 = 0;
@@ -192,7 +191,6 @@ impl TestFileCompiler {
// First, rename the function // First, rename the function
let func_original_name = func.name; let func_original_name = func.name;
func.name = UserFuncName::User(defined_func.new_name.clone()); func.name = UserFuncName::User(defined_func.new_name.clone());
trace!("Rename func from '{}' to '{}'", func_original_name, func.name);
// Rename any functions that it references // Rename any functions that it references
// Do this in stages to appease the borrow checker // Do this in stages to appease the borrow checker
@@ -235,10 +233,6 @@ impl TestFileCompiler {
.get(&func.name) .get(&func.name)
.ok_or(anyhow!("Undeclared function {} found!", &func.name))?; .ok_or(anyhow!("Undeclared function {} found!", &func.name))?;
if log::log_enabled!(log::Level::Trace) {
self.ctx.set_disasm(true);
}
self.ctx.func = self.apply_func_rename(func, defined_func)?; self.ctx.func = self.apply_func_rename(func, defined_func)?;
self.module.define_function_with_control_plane( self.module.define_function_with_control_plane(
defined_func.func_id, defined_func.func_id,
@@ -283,11 +277,6 @@ impl TestFileCompiler {
// available). // available).
self.module.finalize_definitions()?; self.module.finalize_definitions()?;
for (name, func) in self.defined_functions.iter() {
let ptr = self.module.get_finalized_function(func.func_id);
trace!("Function '{}' at {:#X}", name, ptr as u64);
}
Ok(CompiledTestFile { Ok(CompiledTestFile {
module: Some(self.module), module: Some(self.module),
defined_functions: self.defined_functions, defined_functions: self.defined_functions,