This PR switches Cranelift over to the new register allocator, regalloc2. See [this document](https://gist.github.com/cfallin/08553421a91f150254fe878f67301801) for a summary of the design changes. This switchover has implications for core VCode/MachInst types and the lowering pass. Overall, this change brings improvements to both compile time and speed of generated code (runtime), as reported in #3942: ``` Benchmark Compilation (wallclock) Execution (wallclock) blake3-scalar 25% faster 28% faster blake3-simd no diff no diff meshoptimizer 19% faster 17% faster pulldown-cmark 17% faster no diff bz2 15% faster no diff SpiderMonkey, 21% faster 2% faster fib(30) clang.wasm 42% faster N/A ```
343 lines
5.1 KiB
Plaintext
343 lines
5.1 KiB
Plaintext
test compile precise-output
|
|
set unwind_info=false
|
|
target aarch64
|
|
|
|
function %f(i8) -> i64 {
|
|
block0(v0: i8):
|
|
v1 = sextend.i64 v0
|
|
v2 = iconst.i64 42
|
|
v3 = iadd.i64 v2, v1
|
|
return v3
|
|
}
|
|
|
|
; block0:
|
|
; sxtb x4, w0
|
|
; add x0, x4, #42
|
|
; ret
|
|
|
|
function %f2(i8, i64) -> i64 {
|
|
block0(v0: i8, v1: i64):
|
|
v2 = sextend.i64 v0
|
|
v3 = iadd.i64 v2, v1
|
|
return v3
|
|
}
|
|
|
|
; block0:
|
|
; add x0, x1, x0, SXTB
|
|
; ret
|
|
|
|
function %i128_uextend_i64(i64) -> i128 {
|
|
block0(v0: i64):
|
|
v1 = uextend.i128 v0
|
|
return v1
|
|
}
|
|
|
|
; block0:
|
|
; movz x1, #0
|
|
; ret
|
|
|
|
function %i128_sextend_i64(i64) -> i128 {
|
|
block0(v0: i64):
|
|
v1 = sextend.i128 v0
|
|
return v1
|
|
}
|
|
|
|
; block0:
|
|
; asr x1, x0, #63
|
|
; ret
|
|
|
|
function %i128_uextend_i32(i32) -> i128 {
|
|
block0(v0: i32):
|
|
v1 = uextend.i128 v0
|
|
return v1
|
|
}
|
|
|
|
; block0:
|
|
; mov w0, w0
|
|
; movz x1, #0
|
|
; ret
|
|
|
|
function %i128_sextend_i32(i32) -> i128 {
|
|
block0(v0: i32):
|
|
v1 = sextend.i128 v0
|
|
return v1
|
|
}
|
|
|
|
; block0:
|
|
; sxtw x0, w0
|
|
; asr x1, x0, #63
|
|
; ret
|
|
|
|
function %i128_uextend_i16(i16) -> i128 {
|
|
block0(v0: i16):
|
|
v1 = uextend.i128 v0
|
|
return v1
|
|
}
|
|
|
|
; block0:
|
|
; uxth w0, w0
|
|
; movz x1, #0
|
|
; ret
|
|
|
|
function %i128_sextend_i16(i16) -> i128 {
|
|
block0(v0: i16):
|
|
v1 = sextend.i128 v0
|
|
return v1
|
|
}
|
|
|
|
; block0:
|
|
; sxth x0, w0
|
|
; asr x1, x0, #63
|
|
; ret
|
|
|
|
function %i128_uextend_i8(i8) -> i128 {
|
|
block0(v0: i8):
|
|
v1 = uextend.i128 v0
|
|
return v1
|
|
}
|
|
|
|
; block0:
|
|
; uxtb w0, w0
|
|
; movz x1, #0
|
|
; ret
|
|
|
|
function %i128_sextend_i8(i8) -> i128 {
|
|
block0(v0: i8):
|
|
v1 = sextend.i128 v0
|
|
return v1
|
|
}
|
|
|
|
; block0:
|
|
; sxtb x0, w0
|
|
; asr x1, x0, #63
|
|
; ret
|
|
|
|
function %i8x16_uextend_i16(i8x16) -> i16 {
|
|
block0(v0: i8x16):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i16 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; umov w0, v0.b[1]
|
|
; ret
|
|
|
|
function %i8x16_uextend_i32(i8x16) -> i32 {
|
|
block0(v0: i8x16):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i32 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; umov w0, v0.b[1]
|
|
; ret
|
|
|
|
function %i8x16_uextend_i64(i8x16) -> i64 {
|
|
block0(v0: i8x16):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i64 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; umov w0, v0.b[1]
|
|
; ret
|
|
|
|
function %i8x16_uextend_i128(i8x16) -> i128 {
|
|
block0(v0: i8x16):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i128 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; umov w0, v0.b[1]
|
|
; movz x1, #0
|
|
; ret
|
|
|
|
function %i8x16_sextend_i16(i8x16) -> i16 {
|
|
block0(v0: i8x16):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i16 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov w0, v0.b[1]
|
|
; ret
|
|
|
|
function %i8x16_sextend_i32(i8x16) -> i32 {
|
|
block0(v0: i8x16):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i32 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov w0, v0.b[1]
|
|
; ret
|
|
|
|
function %i8x16_sextend_i64(i8x16) -> i64 {
|
|
block0(v0: i8x16):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i64 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov x0, v0.b[1]
|
|
; ret
|
|
|
|
function %i8x16_sextend_i128(i8x16) -> i128 {
|
|
block0(v0: i8x16):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i128 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov x0, v0.b[1]
|
|
; asr x1, x0, #63
|
|
; ret
|
|
|
|
function %i16x8_uextend_i32(i16x8) -> i32 {
|
|
block0(v0: i16x8):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i32 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; umov w0, v0.h[1]
|
|
; ret
|
|
|
|
function %i16x8_uextend_i64(i16x8) -> i64 {
|
|
block0(v0: i16x8):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i64 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; umov w0, v0.h[1]
|
|
; ret
|
|
|
|
function %i16x8_uextend_i128(i16x8) -> i128 {
|
|
block0(v0: i16x8):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i128 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; umov w0, v0.h[1]
|
|
; movz x1, #0
|
|
; ret
|
|
|
|
function %i16x8_sextend_i32(i16x8) -> i32 {
|
|
block0(v0: i16x8):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i32 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov w0, v0.h[1]
|
|
; ret
|
|
|
|
function %i16x8_sextend_i64(i16x8) -> i64 {
|
|
block0(v0: i16x8):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i64 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov x0, v0.h[1]
|
|
; ret
|
|
|
|
function %i16x8_sextend_i128(i16x8) -> i128 {
|
|
block0(v0: i16x8):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i128 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov x0, v0.h[1]
|
|
; asr x1, x0, #63
|
|
; ret
|
|
|
|
function %i32x4_uextend_i64(i32x4) -> i64 {
|
|
block0(v0: i32x4):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i64 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; mov w0, v0.s[1]
|
|
; ret
|
|
|
|
function %i32x4_uextend_i128(i32x4) -> i128 {
|
|
block0(v0: i32x4):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i128 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; mov w0, v0.s[1]
|
|
; movz x1, #0
|
|
; ret
|
|
|
|
function %i32x4_sextend_i64(i32x4) -> i64 {
|
|
block0(v0: i32x4):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i64 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov x0, v0.s[1]
|
|
; ret
|
|
|
|
function %i32x4_sextend_i128(i32x4) -> i128 {
|
|
block0(v0: i32x4):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i128 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; smov x0, v0.s[1]
|
|
; asr x1, x0, #63
|
|
; ret
|
|
|
|
function %i64x2_uextend_i128(i64x2) -> i128 {
|
|
block0(v0: i64x2):
|
|
v1 = extractlane v0, 1
|
|
v2 = uextend.i128 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; mov x0, v0.d[1]
|
|
; movz x1, #0
|
|
; ret
|
|
|
|
function %i64x2_sextend_i128(i64x2) -> i128 {
|
|
block0(v0: i64x2):
|
|
v1 = extractlane v0, 1
|
|
v2 = sextend.i128 v1
|
|
return v2
|
|
}
|
|
|
|
; block0:
|
|
; mov x0, v0.d[1]
|
|
; asr x1, x0, #63
|
|
; ret
|
|
|