Merge pull request #1930 from cfallin/spectre-heap
Spectre mitigation on heap access overflow checks.
This commit is contained in:
@@ -1395,6 +1395,7 @@ fn define_alu(
|
|||||||
let rotr = shared.by_name("rotr");
|
let rotr = shared.by_name("rotr");
|
||||||
let rotr_imm = shared.by_name("rotr_imm");
|
let rotr_imm = shared.by_name("rotr_imm");
|
||||||
let selectif = shared.by_name("selectif");
|
let selectif = shared.by_name("selectif");
|
||||||
|
let selectif_spectre_guard = shared.by_name("selectif_spectre_guard");
|
||||||
let sshr = shared.by_name("sshr");
|
let sshr = shared.by_name("sshr");
|
||||||
let sshr_imm = shared.by_name("sshr_imm");
|
let sshr_imm = shared.by_name("sshr_imm");
|
||||||
let trueff = shared.by_name("trueff");
|
let trueff = shared.by_name("trueff");
|
||||||
@@ -1608,6 +1609,11 @@ fn define_alu(
|
|||||||
|
|
||||||
// Conditional move (a.k.a integer select).
|
// Conditional move (a.k.a integer select).
|
||||||
e.enc_i32_i64(selectif, rec_cmov.opcodes(&CMOV_OVERFLOW));
|
e.enc_i32_i64(selectif, rec_cmov.opcodes(&CMOV_OVERFLOW));
|
||||||
|
// A Spectre-guard integer select is exactly the same as a selectif, but
|
||||||
|
// is not associated with any other legalization rules and is not
|
||||||
|
// recognized by any optimizations, so it must arrive here unmodified
|
||||||
|
// and in its original place.
|
||||||
|
e.enc_i32_i64(selectif_spectre_guard, rec_cmov.opcodes(&CMOV_OVERFLOW));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[inline(never)]
|
#[inline(never)]
|
||||||
|
|||||||
@@ -1748,6 +1748,34 @@ pub(crate) fn define(
|
|||||||
.operands_out(vec![a]),
|
.operands_out(vec![a]),
|
||||||
);
|
);
|
||||||
|
|
||||||
|
ig.push(
|
||||||
|
Inst::new(
|
||||||
|
"selectif_spectre_guard",
|
||||||
|
r#"
|
||||||
|
Conditional select intended for Spectre guards.
|
||||||
|
|
||||||
|
This operation is semantically equivalent to a selectif instruction.
|
||||||
|
However, it is guaranteed to not be removed or otherwise altered by any
|
||||||
|
optimization pass, and is guaranteed to result in a conditional-move
|
||||||
|
instruction, not a branch-based lowering. As such, it is suitable
|
||||||
|
for use when producing Spectre guards. For example, a bounds-check
|
||||||
|
may guard against unsafe speculation past a bounds-check conditional
|
||||||
|
branch by passing the address or index to be accessed through a
|
||||||
|
conditional move, also gated on the same condition. Because no
|
||||||
|
Spectre-vulnerable processors are known to perform speculation on
|
||||||
|
conditional move instructions, this is guaranteed to pick the
|
||||||
|
correct input. If the selected input in case of overflow is a "safe"
|
||||||
|
value, for example a null pointer that causes an exception in the
|
||||||
|
speculative path, this ensures that no Spectre vulnerability will
|
||||||
|
exist.
|
||||||
|
"#,
|
||||||
|
&formats.int_select,
|
||||||
|
)
|
||||||
|
.operands_in(vec![cc, flags, x, y])
|
||||||
|
.operands_out(vec![a])
|
||||||
|
.other_side_effects(true),
|
||||||
|
);
|
||||||
|
|
||||||
let c = &Operand::new("c", Any).with_doc("Controlling value to test");
|
let c = &Operand::new("c", Any).with_doc("Controlling value to test");
|
||||||
ig.push(
|
ig.push(
|
||||||
Inst::new(
|
Inst::new(
|
||||||
|
|||||||
@@ -264,5 +264,23 @@ pub(crate) fn define() -> SettingGroup {
|
|||||||
true,
|
true,
|
||||||
);
|
);
|
||||||
|
|
||||||
|
// Spectre options.
|
||||||
|
|
||||||
|
settings.add_bool(
|
||||||
|
"enable_heap_access_spectre_mitigation",
|
||||||
|
r#"
|
||||||
|
Enable Spectre mitigation on heap bounds checks.
|
||||||
|
|
||||||
|
This is a no-op for any heap that needs no bounds checks; e.g.,
|
||||||
|
if the limit is static and the guard region is large enough that
|
||||||
|
the index cannot reach past it.
|
||||||
|
|
||||||
|
This option is enabled by default because it is highly
|
||||||
|
recommended for secure sandboxing. The embedder should consider
|
||||||
|
the security implications carefully before disabling this option.
|
||||||
|
"#,
|
||||||
|
true,
|
||||||
|
);
|
||||||
|
|
||||||
settings.build()
|
settings.build()
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1023,7 +1023,7 @@ pub(crate) fn lower_insn_to_regs<C: LowerCtx<I = Inst>>(
|
|||||||
// Nothing.
|
// Nothing.
|
||||||
}
|
}
|
||||||
|
|
||||||
Opcode::Select | Opcode::Selectif => {
|
Opcode::Select | Opcode::Selectif | Opcode::SelectifSpectreGuard => {
|
||||||
let cond = if op == Opcode::Select {
|
let cond = if op == Opcode::Select {
|
||||||
let (cmp_op, narrow_mode) = if ty_bits(ctx.input_ty(insn, 0)) > 32 {
|
let (cmp_op, narrow_mode) = if ty_bits(ctx.input_ty(insn, 0)) > 32 {
|
||||||
(ALUOp::SubS64, NarrowValueMode::ZeroExtend64)
|
(ALUOp::SubS64, NarrowValueMode::ZeroExtend64)
|
||||||
|
|||||||
@@ -66,19 +66,14 @@ fn dynamic_addr(
|
|||||||
|
|
||||||
// Start with the bounds check. Trap if `offset + access_size > bound`.
|
// Start with the bounds check. Trap if `offset + access_size > bound`.
|
||||||
let bound = pos.ins().global_value(offset_ty, bound_gv);
|
let bound = pos.ins().global_value(offset_ty, bound_gv);
|
||||||
let oob;
|
let (cc, lhs, bound) = if access_size == 1 {
|
||||||
if access_size == 1 {
|
|
||||||
// `offset > bound - 1` is the same as `offset >= bound`.
|
// `offset > bound - 1` is the same as `offset >= bound`.
|
||||||
oob = pos
|
(IntCC::UnsignedGreaterThanOrEqual, offset, bound)
|
||||||
.ins()
|
|
||||||
.icmp(IntCC::UnsignedGreaterThanOrEqual, offset, bound);
|
|
||||||
} else if access_size <= min_size {
|
} else if access_size <= min_size {
|
||||||
// We know that bound >= min_size, so here we can compare `offset > bound - access_size`
|
// We know that bound >= min_size, so here we can compare `offset > bound - access_size`
|
||||||
// without wrapping.
|
// without wrapping.
|
||||||
let adj_bound = pos.ins().iadd_imm(bound, -(access_size as i64));
|
let adj_bound = pos.ins().iadd_imm(bound, -(access_size as i64));
|
||||||
oob = pos
|
(IntCC::UnsignedGreaterThan, offset, adj_bound)
|
||||||
.ins()
|
|
||||||
.icmp(IntCC::UnsignedGreaterThan, offset, adj_bound);
|
|
||||||
} else {
|
} else {
|
||||||
// We need an overflow check for the adjusted offset.
|
// We need an overflow check for the adjusted offset.
|
||||||
let access_size_val = pos.ins().iconst(offset_ty, access_size as i64);
|
let access_size_val = pos.ins().iconst(offset_ty, access_size as i64);
|
||||||
@@ -88,13 +83,27 @@ fn dynamic_addr(
|
|||||||
overflow,
|
overflow,
|
||||||
ir::TrapCode::HeapOutOfBounds,
|
ir::TrapCode::HeapOutOfBounds,
|
||||||
);
|
);
|
||||||
oob = pos
|
(IntCC::UnsignedGreaterThan, adj_offset, bound)
|
||||||
.ins()
|
};
|
||||||
.icmp(IntCC::UnsignedGreaterThan, adj_offset, bound);
|
let oob = pos.ins().icmp(cc, lhs, bound);
|
||||||
}
|
|
||||||
pos.ins().trapnz(oob, ir::TrapCode::HeapOutOfBounds);
|
pos.ins().trapnz(oob, ir::TrapCode::HeapOutOfBounds);
|
||||||
|
|
||||||
compute_addr(isa, inst, heap, addr_ty, offset, offset_ty, pos.func);
|
let spectre_oob_comparison = if isa.flags().enable_heap_access_spectre_mitigation() {
|
||||||
|
Some((cc, lhs, bound))
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
};
|
||||||
|
|
||||||
|
compute_addr(
|
||||||
|
isa,
|
||||||
|
inst,
|
||||||
|
heap,
|
||||||
|
addr_ty,
|
||||||
|
offset,
|
||||||
|
offset_ty,
|
||||||
|
pos.func,
|
||||||
|
spectre_oob_comparison,
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Expand a `heap_addr` for a static heap.
|
/// Expand a `heap_addr` for a static heap.
|
||||||
@@ -146,20 +155,35 @@ fn static_addr(
|
|||||||
// With that we have an optimization here where with 32-bit offsets and
|
// With that we have an optimization here where with 32-bit offsets and
|
||||||
// `bound - access_size >= 4GB` we can omit a bounds check.
|
// `bound - access_size >= 4GB` we can omit a bounds check.
|
||||||
let limit = bound - access_size;
|
let limit = bound - access_size;
|
||||||
|
let mut spectre_oob_comparison = None;
|
||||||
if offset_ty != ir::types::I32 || limit < 0xffff_ffff {
|
if offset_ty != ir::types::I32 || limit < 0xffff_ffff {
|
||||||
let oob = if limit & 1 == 1 {
|
let (cc, lhs, limit_imm) = if limit & 1 == 1 {
|
||||||
// Prefer testing `offset >= limit - 1` when limit is odd because an even number is
|
// Prefer testing `offset >= limit - 1` when limit is odd because an even number is
|
||||||
// likely to be a convenient constant on ARM and other RISC architectures.
|
// likely to be a convenient constant on ARM and other RISC architectures.
|
||||||
pos.ins()
|
let limit = limit as i64 - 1;
|
||||||
.icmp_imm(IntCC::UnsignedGreaterThanOrEqual, offset, limit as i64 - 1)
|
(IntCC::UnsignedGreaterThanOrEqual, offset, limit)
|
||||||
} else {
|
} else {
|
||||||
pos.ins()
|
let limit = limit as i64;
|
||||||
.icmp_imm(IntCC::UnsignedGreaterThan, offset, limit as i64)
|
(IntCC::UnsignedGreaterThan, offset, limit)
|
||||||
};
|
};
|
||||||
|
let oob = pos.ins().icmp_imm(cc, lhs, limit_imm);
|
||||||
pos.ins().trapnz(oob, ir::TrapCode::HeapOutOfBounds);
|
pos.ins().trapnz(oob, ir::TrapCode::HeapOutOfBounds);
|
||||||
|
if isa.flags().enable_heap_access_spectre_mitigation() {
|
||||||
|
let limit = pos.ins().iconst(offset_ty, limit_imm);
|
||||||
|
spectre_oob_comparison = Some((cc, lhs, limit));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
compute_addr(isa, inst, heap, addr_ty, offset, offset_ty, pos.func);
|
compute_addr(
|
||||||
|
isa,
|
||||||
|
inst,
|
||||||
|
heap,
|
||||||
|
addr_ty,
|
||||||
|
offset,
|
||||||
|
offset_ty,
|
||||||
|
pos.func,
|
||||||
|
spectre_oob_comparison,
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Emit code for the base address computation of a `heap_addr` instruction.
|
/// Emit code for the base address computation of a `heap_addr` instruction.
|
||||||
@@ -171,6 +195,11 @@ fn compute_addr(
|
|||||||
mut offset: ir::Value,
|
mut offset: ir::Value,
|
||||||
offset_ty: ir::Type,
|
offset_ty: ir::Type,
|
||||||
func: &mut ir::Function,
|
func: &mut ir::Function,
|
||||||
|
// If we are performing Spectre mitigation with conditional selects, the
|
||||||
|
// values to compare and the condition code that indicates an out-of bounds
|
||||||
|
// condition; on this condition, the conditional move will choose a
|
||||||
|
// speculatively safe address (a zero / null pointer) instead.
|
||||||
|
spectre_oob_comparison: Option<(IntCC, ir::Value, ir::Value)>,
|
||||||
) {
|
) {
|
||||||
let mut pos = FuncCursor::new(func).at_inst(inst);
|
let mut pos = FuncCursor::new(func).at_inst(inst);
|
||||||
pos.use_srcloc(inst);
|
pos.use_srcloc(inst);
|
||||||
@@ -198,5 +227,15 @@ fn compute_addr(
|
|||||||
pos.ins().global_value(addr_ty, base_gv)
|
pos.ins().global_value(addr_ty, base_gv)
|
||||||
};
|
};
|
||||||
|
|
||||||
pos.func.dfg.replace(inst).iadd(base, offset);
|
if let Some((cc, a, b)) = spectre_oob_comparison {
|
||||||
|
let final_addr = pos.ins().iadd(base, offset);
|
||||||
|
let zero = pos.ins().iconst(addr_ty, 0);
|
||||||
|
let flags = pos.ins().ifcmp(a, b);
|
||||||
|
pos.func
|
||||||
|
.dfg
|
||||||
|
.replace(inst)
|
||||||
|
.selectif_spectre_guard(addr_ty, cc, flags, zero, final_addr);
|
||||||
|
} else {
|
||||||
|
pos.func.dfg.replace(inst).iadd(base, offset);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -399,6 +399,7 @@ emit_all_ones_funcaddrs = false
|
|||||||
enable_probestack = true
|
enable_probestack = true
|
||||||
probestack_func_adjusts_sp = false
|
probestack_func_adjusts_sp = false
|
||||||
enable_jump_tables = true
|
enable_jump_tables = true
|
||||||
|
enable_heap_access_spectre_mitigation = true
|
||||||
"#
|
"#
|
||||||
);
|
);
|
||||||
assert_eq!(f.opt_level(), super::OptLevel::None);
|
assert_eq!(f.opt_level(), super::OptLevel::None);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
test legalizer
|
test legalizer
|
||||||
|
set enable_heap_access_spectre_mitigation=false
|
||||||
target x86_64
|
target x86_64
|
||||||
|
|
||||||
; Test legalization for various forms of heap addresses.
|
; Test legalization for various forms of heap addresses.
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
; Test the legalization of memory objects.
|
; Test the legalization of memory objects.
|
||||||
test legalizer
|
test legalizer
|
||||||
|
set enable_heap_access_spectre_mitigation=false
|
||||||
target x86_64
|
target x86_64
|
||||||
|
|
||||||
; regex: V=v\d+
|
; regex: V=v\d+
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
test compile
|
test compile
|
||||||
|
set enable_heap_access_spectre_mitigation=true
|
||||||
target aarch64
|
target aarch64
|
||||||
|
|
||||||
function %dynamic_heap_check(i64 vmctx, i32) -> i64 {
|
function %dynamic_heap_check(i64 vmctx, i32) -> i64 {
|
||||||
@@ -11,20 +12,23 @@ block0(v0: i64, v1: i32):
|
|||||||
return v2
|
return v2
|
||||||
}
|
}
|
||||||
|
|
||||||
; check: stp fp, lr, [sp, #-16]!
|
; check: Block 0:
|
||||||
; nextln: mov fp, sp
|
; check: stp fp, lr, [sp, #-16]!
|
||||||
; nextln: ldur w2, [x0]
|
; nextln: mov fp, sp
|
||||||
; nextln: add w2, w2, #0
|
; nextln: ldur w2, [x0]
|
||||||
; nextln: subs wzr, w1, w2
|
; nextln: add w2, w2, #0
|
||||||
; nextln: b.ls label1 ; b label2
|
; nextln: subs wzr, w1, w2
|
||||||
; nextln: Block 1:
|
; nextln: b.ls label1 ; b label2
|
||||||
; check: add x0, x0, x1, UXTW
|
; check: Block 1:
|
||||||
; nextln: mov sp, fp
|
; check: add x0, x0, x1, UXTW
|
||||||
; nextln: ldp fp, lr, [sp], #16
|
; nextln: subs wzr, w1, w2
|
||||||
; nextln: ret
|
; nextln: movz x1, #0
|
||||||
; nextln: Block 2:
|
; nextln: csel x0, x1, x0, hi
|
||||||
; check: udf
|
; nextln: mov sp, fp
|
||||||
|
; nextln: ldp fp, lr, [sp], #16
|
||||||
|
; nextln: ret
|
||||||
|
; check: Block 2:
|
||||||
|
; check: udf
|
||||||
|
|
||||||
function %static_heap_check(i64 vmctx, i32) -> i64 {
|
function %static_heap_check(i64 vmctx, i32) -> i64 {
|
||||||
gv0 = vmctx
|
gv0 = vmctx
|
||||||
@@ -35,15 +39,18 @@ block0(v0: i64, v1: i32):
|
|||||||
return v2
|
return v2
|
||||||
}
|
}
|
||||||
|
|
||||||
; check: stp fp, lr, [sp, #-16]!
|
; check: Block 0:
|
||||||
; nextln: mov fp, sp
|
; check: stp fp, lr, [sp, #-16]!
|
||||||
; nextln: subs wzr, w1, #65536
|
; nextln: mov fp, sp
|
||||||
; nextln: b.ls label1 ; b label2
|
; nextln: subs wzr, w1, #65536
|
||||||
; nextln: Block 1:
|
; nextln: b.ls label1 ; b label2
|
||||||
; check: add x0, x0, x1, UXTW
|
; check: Block 1:
|
||||||
; nextln: mov sp, fp
|
; check: add x0, x0, x1, UXTW
|
||||||
; nextln: ldp fp, lr, [sp], #16
|
; nextln: subs wzr, w1, #65536
|
||||||
; nextln: ret
|
; nextln: movz x1, #0
|
||||||
; nextln: Block 2:
|
; nextln: csel x0, x1, x0, hi
|
||||||
; check: udf
|
; nextln: mov sp, fp
|
||||||
|
; nextln: ldp fp, lr, [sp], #16
|
||||||
|
; nextln: ret
|
||||||
|
; check: Block 2:
|
||||||
|
; check: udf
|
||||||
|
|||||||
Reference in New Issue
Block a user