MachInst backends: handle SourceLocs out-of-band, not in Insts.
In existing MachInst backends, many instructions -- any that can trap or result in a relocation -- carry `SourceLoc` values in order to propagate the location-in-original-source to use to describe resulting traps or relocation errors. This is quite tedious, and also error-prone: it is likely that the necessary plumbing will be missed in some cases, and in any case, it's unnecessarily verbose. This PR factors out the `SourceLoc` handling so that it is tracked during emission as part of the `EmitState`, and plumbed through automatically by the machine-independent framework. Instruction emission code that directly emits trap or relocation records can query the current location as necessary. Then we only need to ensure that memory references and trap instructions, at their (one) emission point rather than their (many) lowering/generation points, are wired up correctly. This does have the side-effect that some loads and stores that do not correspond directly to user code's heap accesses will have unnecessary but harmless trap metadata. For example, the load that fetches a code offset from a jump table will have a 'heap out of bounds' trap record attached to it; but because it is bounds-checked, and will never actually trap if the lowering is correct, this should be harmless. The simplicity improvement here seemed more worthwhile to me than plumbing through a "corresponds to user-level load/store" bit, because the latter is a bit complex when we allow for op merging. Closes #2290: though it does not implement a full "metadata" scheme as described in that issue, this seems simpler overall.
This commit is contained in:
@@ -484,9 +484,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
// msub rd, rd, rm, rn ; rd = rn - rd * rm
|
||||
|
||||
// Check for divide by 0.
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::IntegerDivisionByZero);
|
||||
let trap_code = TrapCode::IntegerDivisionByZero;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Zero(rm),
|
||||
});
|
||||
|
||||
@@ -507,9 +507,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
// udf ; signed overflow
|
||||
|
||||
// Check for divide by 0.
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::IntegerDivisionByZero);
|
||||
let trap_code = TrapCode::IntegerDivisionByZero;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Zero(rm),
|
||||
});
|
||||
|
||||
@@ -535,9 +535,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
nzcv: NZCV::new(false, false, false, false),
|
||||
cond: Cond::Eq,
|
||||
});
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::IntegerOverflow);
|
||||
let trap_code = TrapCode::IntegerOverflow;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Cond(Cond::Vs),
|
||||
});
|
||||
} else {
|
||||
@@ -545,9 +545,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
// udf ; divide by zero
|
||||
|
||||
// Check for divide by 0.
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::IntegerDivisionByZero);
|
||||
let trap_code = TrapCode::IntegerDivisionByZero;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Zero(rm),
|
||||
});
|
||||
}
|
||||
@@ -1161,27 +1161,20 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let mem = lower_address(ctx, elem_ty, &inputs[..], off);
|
||||
let rd = get_output_reg(ctx, outputs[0]);
|
||||
|
||||
let memflags = ctx.memflags(insn).expect("memory flags");
|
||||
let srcloc = if !memflags.notrap() {
|
||||
Some(ctx.srcloc(insn))
|
||||
} else {
|
||||
None
|
||||
};
|
||||
|
||||
ctx.emit(match (ty_bits(elem_ty), sign_extend, is_float) {
|
||||
(1, _, _) => Inst::ULoad8 { rd, mem, srcloc },
|
||||
(8, false, _) => Inst::ULoad8 { rd, mem, srcloc },
|
||||
(8, true, _) => Inst::SLoad8 { rd, mem, srcloc },
|
||||
(16, false, _) => Inst::ULoad16 { rd, mem, srcloc },
|
||||
(16, true, _) => Inst::SLoad16 { rd, mem, srcloc },
|
||||
(32, false, false) => Inst::ULoad32 { rd, mem, srcloc },
|
||||
(32, true, false) => Inst::SLoad32 { rd, mem, srcloc },
|
||||
(32, _, true) => Inst::FpuLoad32 { rd, mem, srcloc },
|
||||
(64, _, false) => Inst::ULoad64 { rd, mem, srcloc },
|
||||
(1, _, _) => Inst::ULoad8 { rd, mem },
|
||||
(8, false, _) => Inst::ULoad8 { rd, mem },
|
||||
(8, true, _) => Inst::SLoad8 { rd, mem },
|
||||
(16, false, _) => Inst::ULoad16 { rd, mem },
|
||||
(16, true, _) => Inst::SLoad16 { rd, mem },
|
||||
(32, false, false) => Inst::ULoad32 { rd, mem },
|
||||
(32, true, false) => Inst::SLoad32 { rd, mem },
|
||||
(32, _, true) => Inst::FpuLoad32 { rd, mem },
|
||||
(64, _, false) => Inst::ULoad64 { rd, mem },
|
||||
// Note that we treat some of the vector loads as scalar floating-point loads,
|
||||
// which is correct in a little endian environment.
|
||||
(64, _, true) => Inst::FpuLoad64 { rd, mem, srcloc },
|
||||
(128, _, _) => Inst::FpuLoad128 { rd, mem, srcloc },
|
||||
(64, _, true) => Inst::FpuLoad64 { rd, mem },
|
||||
(128, _, _) => Inst::FpuLoad128 { rd, mem },
|
||||
_ => panic!("Unsupported size in load"),
|
||||
});
|
||||
|
||||
@@ -1209,14 +1202,8 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let off = ctx.data(insn).load_store_offset().unwrap();
|
||||
let ty = ty.unwrap();
|
||||
let mem = lower_address(ctx, ty.lane_type(), &inputs[..], off);
|
||||
let memflags = ctx.memflags(insn).expect("memory flags");
|
||||
let rd = get_output_reg(ctx, outputs[0]);
|
||||
let size = VectorSize::from_ty(ty);
|
||||
let srcloc = if memflags.notrap() {
|
||||
None
|
||||
} else {
|
||||
Some(ctx.srcloc(insn))
|
||||
};
|
||||
let tmp = ctx.alloc_tmp(RegClass::I64, I64);
|
||||
|
||||
ctx.emit(Inst::LoadAddr { rd: tmp, mem });
|
||||
@@ -1224,7 +1211,6 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
rd,
|
||||
rn: tmp.to_reg(),
|
||||
size,
|
||||
srcloc,
|
||||
});
|
||||
}
|
||||
|
||||
@@ -1249,21 +1235,14 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let mem = lower_address(ctx, elem_ty, &inputs[1..], off);
|
||||
let rd = put_input_in_reg(ctx, inputs[0], NarrowValueMode::None);
|
||||
|
||||
let memflags = ctx.memflags(insn).expect("memory flags");
|
||||
let srcloc = if !memflags.notrap() {
|
||||
Some(ctx.srcloc(insn))
|
||||
} else {
|
||||
None
|
||||
};
|
||||
|
||||
ctx.emit(match (ty_bits(elem_ty), is_float) {
|
||||
(1, _) | (8, _) => Inst::Store8 { rd, mem, srcloc },
|
||||
(16, _) => Inst::Store16 { rd, mem, srcloc },
|
||||
(32, false) => Inst::Store32 { rd, mem, srcloc },
|
||||
(32, true) => Inst::FpuStore32 { rd, mem, srcloc },
|
||||
(64, false) => Inst::Store64 { rd, mem, srcloc },
|
||||
(64, true) => Inst::FpuStore64 { rd, mem, srcloc },
|
||||
(128, _) => Inst::FpuStore128 { rd, mem, srcloc },
|
||||
(1, _) | (8, _) => Inst::Store8 { rd, mem },
|
||||
(16, _) => Inst::Store16 { rd, mem },
|
||||
(32, false) => Inst::Store32 { rd, mem },
|
||||
(32, true) => Inst::FpuStore32 { rd, mem },
|
||||
(64, false) => Inst::Store64 { rd, mem },
|
||||
(64, true) => Inst::FpuStore64 { rd, mem },
|
||||
(128, _) => Inst::FpuStore128 { rd, mem },
|
||||
_ => panic!("Unsupported size in store"),
|
||||
});
|
||||
}
|
||||
@@ -1291,12 +1270,6 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let mut r_arg2 = put_input_in_reg(ctx, inputs[1], NarrowValueMode::None);
|
||||
let ty_access = ty.unwrap();
|
||||
assert!(is_valid_atomic_transaction_ty(ty_access));
|
||||
let memflags = ctx.memflags(insn).expect("memory flags");
|
||||
let srcloc = if !memflags.notrap() {
|
||||
Some(ctx.srcloc(insn))
|
||||
} else {
|
||||
None
|
||||
};
|
||||
// Make sure that both args are in virtual regs, since in effect
|
||||
// we have to do a parallel copy to get them safely to the AtomicRMW input
|
||||
// regs, and that's not guaranteed safe if either is in a real reg.
|
||||
@@ -1307,11 +1280,7 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
ctx.emit(Inst::gen_move(Writable::from_reg(xreg(26)), r_arg2, I64));
|
||||
// Now the AtomicRMW insn itself
|
||||
let op = inst_common::AtomicRmwOp::from(ctx.data(insn).atomic_rmw_op().unwrap());
|
||||
ctx.emit(Inst::AtomicRMW {
|
||||
ty: ty_access,
|
||||
op,
|
||||
srcloc,
|
||||
});
|
||||
ctx.emit(Inst::AtomicRMW { ty: ty_access, op });
|
||||
// And finally, copy the preordained AtomicRMW output reg to its destination.
|
||||
ctx.emit(Inst::gen_move(r_dst, xreg(27), I64));
|
||||
// Also, x24 and x28 are trashed. `fn aarch64_get_regs` must mention that.
|
||||
@@ -1327,12 +1296,6 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let mut r_replacement = put_input_in_reg(ctx, inputs[2], NarrowValueMode::None);
|
||||
let ty_access = ty.unwrap();
|
||||
assert!(is_valid_atomic_transaction_ty(ty_access));
|
||||
let memflags = ctx.memflags(insn).expect("memory flags");
|
||||
let srcloc = if !memflags.notrap() {
|
||||
Some(ctx.srcloc(insn))
|
||||
} else {
|
||||
None
|
||||
};
|
||||
// Make sure that all three args are in virtual regs. See corresponding comment
|
||||
// for `Opcode::AtomicRmw` above.
|
||||
r_addr = ctx.ensure_in_vreg(r_addr, I64);
|
||||
@@ -1351,10 +1314,7 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
I64,
|
||||
));
|
||||
// Now the AtomicCAS itself, implemented in the normal way, with an LL-SC loop
|
||||
ctx.emit(Inst::AtomicCAS {
|
||||
ty: ty_access,
|
||||
srcloc,
|
||||
});
|
||||
ctx.emit(Inst::AtomicCAS { ty: ty_access });
|
||||
// And finally, copy the preordained AtomicCAS output reg to its destination.
|
||||
ctx.emit(Inst::gen_move(r_dst, xreg(27), I64));
|
||||
// Also, x24 and x28 are trashed. `fn aarch64_get_regs` must mention that.
|
||||
@@ -1365,17 +1325,10 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let r_addr = put_input_in_reg(ctx, inputs[0], NarrowValueMode::None);
|
||||
let ty_access = ty.unwrap();
|
||||
assert!(is_valid_atomic_transaction_ty(ty_access));
|
||||
let memflags = ctx.memflags(insn).expect("memory flags");
|
||||
let srcloc = if !memflags.notrap() {
|
||||
Some(ctx.srcloc(insn))
|
||||
} else {
|
||||
None
|
||||
};
|
||||
ctx.emit(Inst::AtomicLoad {
|
||||
ty: ty_access,
|
||||
r_data,
|
||||
r_addr,
|
||||
srcloc,
|
||||
});
|
||||
}
|
||||
|
||||
@@ -1384,17 +1337,10 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let r_addr = put_input_in_reg(ctx, inputs[1], NarrowValueMode::None);
|
||||
let ty_access = ctx.input_ty(insn, 0);
|
||||
assert!(is_valid_atomic_transaction_ty(ty_access));
|
||||
let memflags = ctx.memflags(insn).expect("memory flags");
|
||||
let srcloc = if !memflags.notrap() {
|
||||
Some(ctx.srcloc(insn))
|
||||
} else {
|
||||
None
|
||||
};
|
||||
ctx.emit(Inst::AtomicStore {
|
||||
ty: ty_access,
|
||||
r_data,
|
||||
r_addr,
|
||||
srcloc,
|
||||
});
|
||||
}
|
||||
|
||||
@@ -1811,12 +1757,12 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
}
|
||||
|
||||
Opcode::Trap | Opcode::ResumableTrap => {
|
||||
let trap_info = (ctx.srcloc(insn), ctx.data(insn).trap_code().unwrap());
|
||||
ctx.emit_safepoint(Inst::Udf { trap_info });
|
||||
let trap_code = ctx.data(insn).trap_code().unwrap();
|
||||
ctx.emit_safepoint(Inst::Udf { trap_code });
|
||||
}
|
||||
|
||||
Opcode::Trapif | Opcode::Trapff => {
|
||||
let trap_info = (ctx.srcloc(insn), ctx.data(insn).trap_code().unwrap());
|
||||
let trap_code = ctx.data(insn).trap_code().unwrap();
|
||||
|
||||
let cond = if maybe_input_insn(ctx, inputs[0], Opcode::IaddIfcout).is_some() {
|
||||
let condcode = ctx.data(insn).cond_code().unwrap();
|
||||
@@ -1847,7 +1793,7 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
};
|
||||
|
||||
ctx.emit_safepoint(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Cond(cond),
|
||||
});
|
||||
}
|
||||
@@ -1864,11 +1810,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let rd = get_output_reg(ctx, outputs[0]);
|
||||
let (extname, _) = ctx.call_target(insn).unwrap();
|
||||
let extname = extname.clone();
|
||||
let loc = ctx.srcloc(insn);
|
||||
ctx.emit(Inst::LoadExtName {
|
||||
rd,
|
||||
name: Box::new(extname),
|
||||
srcloc: loc,
|
||||
offset: 0,
|
||||
});
|
||||
}
|
||||
@@ -1881,17 +1825,14 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
let rd = get_output_reg(ctx, outputs[0]);
|
||||
let (extname, _, offset) = ctx.symbol_value(insn).unwrap();
|
||||
let extname = extname.clone();
|
||||
let loc = ctx.srcloc(insn);
|
||||
ctx.emit(Inst::LoadExtName {
|
||||
rd,
|
||||
name: Box::new(extname),
|
||||
srcloc: loc,
|
||||
offset,
|
||||
});
|
||||
}
|
||||
|
||||
Opcode::Call | Opcode::CallIndirect => {
|
||||
let loc = ctx.srcloc(insn);
|
||||
let caller_conv = ctx.abi().call_conv();
|
||||
let (mut abi, inputs) = match op {
|
||||
Opcode::Call => {
|
||||
@@ -1901,7 +1842,7 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
assert!(inputs.len() == sig.params.len());
|
||||
assert!(outputs.len() == sig.returns.len());
|
||||
(
|
||||
AArch64ABICaller::from_func(sig, &extname, dist, loc, caller_conv)?,
|
||||
AArch64ABICaller::from_func(sig, &extname, dist, caller_conv)?,
|
||||
&inputs[..],
|
||||
)
|
||||
}
|
||||
@@ -1911,7 +1852,7 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
assert!(inputs.len() - 1 == sig.params.len());
|
||||
assert!(outputs.len() == sig.returns.len());
|
||||
(
|
||||
AArch64ABICaller::from_ptr(sig, ptr, loc, op, caller_conv)?,
|
||||
AArch64ABICaller::from_ptr(sig, ptr, op, caller_conv)?,
|
||||
&inputs[1..],
|
||||
)
|
||||
}
|
||||
@@ -2687,9 +2628,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
} else {
|
||||
ctx.emit(Inst::FpuCmp64 { rn, rm: rn });
|
||||
}
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::BadConversionToInteger);
|
||||
let trap_code = TrapCode::BadConversionToInteger;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Cond(lower_fp_condcode(FloatCC::Unordered)),
|
||||
});
|
||||
|
||||
@@ -2739,9 +2680,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
rn,
|
||||
rm: tmp.to_reg(),
|
||||
});
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::IntegerOverflow);
|
||||
let trap_code = TrapCode::IntegerOverflow;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Cond(lower_fp_condcode(low_cond).invert()),
|
||||
});
|
||||
|
||||
@@ -2751,9 +2692,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
rn,
|
||||
rm: tmp.to_reg(),
|
||||
});
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::IntegerOverflow);
|
||||
let trap_code = TrapCode::IntegerOverflow;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Cond(lower_fp_condcode(FloatCC::LessThan).invert()),
|
||||
});
|
||||
} else {
|
||||
@@ -2792,9 +2733,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
rn,
|
||||
rm: tmp.to_reg(),
|
||||
});
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::IntegerOverflow);
|
||||
let trap_code = TrapCode::IntegerOverflow;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Cond(lower_fp_condcode(low_cond).invert()),
|
||||
});
|
||||
|
||||
@@ -2804,9 +2745,9 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
||||
rn,
|
||||
rm: tmp.to_reg(),
|
||||
});
|
||||
let trap_info = (ctx.srcloc(insn), TrapCode::IntegerOverflow);
|
||||
let trap_code = TrapCode::IntegerOverflow;
|
||||
ctx.emit(Inst::TrapIf {
|
||||
trap_info,
|
||||
trap_code,
|
||||
kind: CondBrKind::Cond(lower_fp_condcode(FloatCC::LessThan).invert()),
|
||||
});
|
||||
};
|
||||
|
||||
Reference in New Issue
Block a user