diff options
author | Juergen Ributzka <juergen@apple.com> | 2014-10-07 03:39:59 +0000 |
---|---|---|
committer | Juergen Ributzka <juergen@apple.com> | 2014-10-07 03:39:59 +0000 |
commit | 42bf665f2bcb23525f07d054db001b3bcd3466bb (patch) | |
tree | 72f8e3b380e1b487694611e4052d56e55d5f517b /llvm/test/CodeGen/AArch64/fast-isel-int-ext.ll | |
parent | cb69583c507ca6f41072f5f52177bb3b6ad359cf (diff) | |
download | bcm5719-llvm-42bf665f2bcb23525f07d054db001b3bcd3466bb.tar.gz bcm5719-llvm-42bf665f2bcb23525f07d054db001b3bcd3466bb.zip |
[FastISel][AArch64] Fix "Fold sign-/zero-extends into the load instruction."
This commit fixes an issue with sign-/zero-extending loads that was discovered
by Richard Barton.
We use now the correct load instructions for sign-extending loads to 64bit. Also
updated and added more unit tests.
llvm-svn: 219185
Diffstat (limited to 'llvm/test/CodeGen/AArch64/fast-isel-int-ext.ll')
-rw-r--r-- | llvm/test/CodeGen/AArch64/fast-isel-int-ext.ll | 453 |
1 files changed, 382 insertions, 71 deletions
diff --git a/llvm/test/CodeGen/AArch64/fast-isel-int-ext.ll b/llvm/test/CodeGen/AArch64/fast-isel-int-ext.ll index 115b96d7806..31372ee35bd 100644 --- a/llvm/test/CodeGen/AArch64/fast-isel-int-ext.ll +++ b/llvm/test/CodeGen/AArch64/fast-isel-int-ext.ll @@ -6,9 +6,9 @@ ; ; SHIFT ; -define i64 @load_addr_shift_zext1(i32 zeroext %a, i64 %b) { +define i64 @load_addr_shift_zext1(i32 %a, i64 %b) { ; CHECK-LABEL: load_addr_shift_zext1 -; CHECK: ldr {{x[0-9]+}}, [x1, x0, lsl #3] +; CHECK: ldr {{x[0-9]+}}, [x1, w0, uxtw #3] %1 = zext i32 %a to i64 %2 = shl i64 %1, 3 %3 = add i64 %b, %2 @@ -17,9 +17,9 @@ define i64 @load_addr_shift_zext1(i32 zeroext %a, i64 %b) { ret i64 %5 } -define i64 @load_addr_shift_zext2(i32 signext %a, i64 %b) { +define i64 @load_addr_shift_zext2(i32 zeroext %a, i64 %b) { ; CHECK-LABEL: load_addr_shift_zext2 -; CHECK: ldr {{x[0-9]+}}, [x1, w0, uxtw #3{{\]}} +; CHECK: ldr {{x[0-9]+}}, [x1, x0, lsl #3] %1 = zext i32 %a to i64 %2 = shl i64 %1, 3 %3 = add i64 %b, %2 @@ -28,9 +28,20 @@ define i64 @load_addr_shift_zext2(i32 signext %a, i64 %b) { ret i64 %5 } -define i64 @load_addr_shift_sext1(i32 signext %a, i64 %b) { +define i64 @load_addr_shift_zext3(i32 signext %a, i64 %b) { +; CHECK-LABEL: load_addr_shift_zext3 +; CHECK: ldr {{x[0-9]+}}, [x1, w0, uxtw #3] + %1 = zext i32 %a to i64 + %2 = shl i64 %1, 3 + %3 = add i64 %b, %2 + %4 = inttoptr i64 %3 to i64* + %5 = load i64* %4 + ret i64 %5 +} + +define i64 @load_addr_shift_sext1(i32 %a, i64 %b) { ; CHECK-LABEL: load_addr_shift_sext1 -; CHECK: ldr {{x[0-9]+}}, [x1, x0, lsl #3] +; CHECK: ldr {{x[0-9]+}}, [x1, w0, sxtw #3] %1 = sext i32 %a to i64 %2 = shl i64 %1, 3 %3 = add i64 %b, %2 @@ -50,12 +61,23 @@ define i64 @load_addr_shift_sext2(i32 zeroext %a, i64 %b) { ret i64 %5 } +define i64 @load_addr_shift_sext3(i32 signext %a, i64 %b) { +; CHECK-LABEL: load_addr_shift_sext3 +; CHECK: ldr {{x[0-9]+}}, [x1, x0, lsl #3] + %1 = sext i32 %a to i64 + %2 = shl i64 %1, 3 + %3 = add i64 %b, %2 + %4 = inttoptr i64 %3 to i64* + %5 = load i64* %4 + ret i64 %5 +} + ; ; MUL ; -define i64 @load_addr_mul_zext1(i32 zeroext %a, i64 %b) { +define i64 @load_addr_mul_zext1(i32 %a, i64 %b) { ; CHECK-LABEL: load_addr_mul_zext1 -; CHECK: ldr {{x[0-9]+}}, [x1, x0, lsl #3] +; CHECK: ldr {{x[0-9]+}}, [x1, w0, uxtw #3] %1 = zext i32 %a to i64 %2 = mul i64 %1, 8 %3 = add i64 %b, %2 @@ -64,8 +86,19 @@ define i64 @load_addr_mul_zext1(i32 zeroext %a, i64 %b) { ret i64 %5 } -define i64 @load_addr_mul_zext2(i32 signext %a, i64 %b) { +define i64 @load_addr_mul_zext2(i32 zeroext %a, i64 %b) { ; CHECK-LABEL: load_addr_mul_zext2 +; CHECK: ldr {{x[0-9]+}}, [x1, x0, lsl #3] + %1 = zext i32 %a to i64 + %2 = mul i64 %1, 8 + %3 = add i64 %b, %2 + %4 = inttoptr i64 %3 to i64* + %5 = load i64* %4 + ret i64 %5 +} + +define i64 @load_addr_mul_zext3(i32 signext %a, i64 %b) { +; CHECK-LABEL: load_addr_mul_zext3 ; CHECK: ldr {{x[0-9]+}}, [x1, w0, uxtw #3] %1 = zext i32 %a to i64 %2 = mul i64 %1, 8 @@ -75,9 +108,9 @@ define i64 @load_addr_mul_zext2(i32 signext %a, i64 %b) { ret i64 %5 } -define i64 @load_addr_mul_sext1(i32 signext %a, i64 %b) { +define i64 @load_addr_mul_sext1(i32 %a, i64 %b) { ; CHECK-LABEL: load_addr_mul_sext1 -; CHECK: ldr {{x[0-9]+}}, [x1, x0, lsl #3] +; CHECK: ldr {{x[0-9]+}}, [x1, w0, sxtw #3] %1 = sext i32 %a to i64 %2 = mul i64 %1, 8 %3 = add i64 %b, %2 @@ -97,94 +130,372 @@ define i64 @load_addr_mul_sext2(i32 zeroext %a, i64 %b) { ret i64 %5 } +define i64 @load_addr_mul_sext3(i32 signext %a, i64 %b) { +; CHECK-LABEL: load_addr_mul_sext3 +; CHECK: ldr {{x[0-9]+}}, [x1, x0, lsl #3] + %1 = sext i32 %a to i64 + %2 = mul i64 %1, 8 + %3 = add i64 %b, %2 + %4 = inttoptr i64 %3 to i64* + %5 = load i64* %4 + ret i64 %5 +} + + +; ; Test folding of the sign-/zero-extend into the load instruction. -define i32 @load_zext_i8_to_i32(i8* %a) { -; CHECK-LABEL: load_zext_i8_to_i32 -; CHECK: ldrb w0, [x0] +; + +; Unscaled +define i32 @load_unscaled_zext_i8_to_i32(i64 %a) { +; CHECK-LABEL: load_unscaled_zext_i8_to_i32 +; CHECK: ldurb w0, [x0, #-8] ; CHECK-NOT: uxtb - %1 = load i8* %a - %2 = zext i8 %1 to i32 - ret i32 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i8* + %3 = load i8* %2 + %4 = zext i8 %3 to i32 + ret i32 %4 } -define i32 @load_zext_i16_to_i32(i16* %a) { -; CHECK-LABEL: load_zext_i16_to_i32 -; CHECK: ldrh w0, [x0] +define i32 @load_unscaled_zext_i16_to_i32(i64 %a) { +; CHECK-LABEL: load_unscaled_zext_i16_to_i32 +; CHECK: ldurh w0, [x0, #-8] ; CHECK-NOT: uxth - %1 = load i16* %a - %2 = zext i16 %1 to i32 - ret i32 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i16* + %3 = load i16* %2 + %4 = zext i16 %3 to i32 + ret i32 %4 } -define i64 @load_zext_i8_to_i64(i8* %a) { -; CHECK-LABEL: load_zext_i8_to_i64 -; CHECK: ldrb w0, [x0] +define i64 @load_unscaled_zext_i8_to_i64(i64 %a) { +; CHECK-LABEL: load_unscaled_zext_i8_to_i64 +; CHECK: ldurb w0, [x0, #-8] ; CHECK-NOT: uxtb - %1 = load i8* %a - %2 = zext i8 %1 to i64 - ret i64 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i8* + %3 = load i8* %2 + %4 = zext i8 %3 to i64 + ret i64 %4 } -define i64 @load_zext_i16_to_i64(i16* %a) { -; CHECK-LABEL: load_zext_i16_to_i64 -; CHECK: ldrh w0, [x0] +define i64 @load_unscaled_zext_i16_to_i64(i64 %a) { +; CHECK-LABEL: load_unscaled_zext_i16_to_i64 +; CHECK: ldurh w0, [x0, #-8] ; CHECK-NOT: uxth - %1 = load i16* %a - %2 = zext i16 %1 to i64 - ret i64 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i16* + %3 = load i16* %2 + %4 = zext i16 %3 to i64 + ret i64 %4 } -define i64 @load_zext_i32_to_i64(i32* %a) { -; CHECK-LABEL: load_zext_i32_to_i64 -; CHECK: ldr w0, [x0] +define i64 @load_unscaled_zext_i32_to_i64(i64 %a) { +; CHECK-LABEL: load_unscaled_zext_i32_to_i64 +; CHECK: ldur w0, [x0, #-8] ; CHECK-NOT: uxtw - %1 = load i32* %a - %2 = zext i32 %1 to i64 - ret i64 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i32* + %3 = load i32* %2 + %4 = zext i32 %3 to i64 + ret i64 %4 } -define i32 @load_sext_i8_to_i32(i8* %a) { -; CHECK-LABEL: load_sext_i8_to_i32 -; CHECK: ldrsb w0, [x0] +define i32 @load_unscaled_sext_i8_to_i32(i64 %a) { +; CHECK-LABEL: load_unscaled_sext_i8_to_i32 +; CHECK: ldursb w0, [x0, #-8] ; CHECK-NOT: sxtb - %1 = load i8* %a - %2 = sext i8 %1 to i32 - ret i32 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i8* + %3 = load i8* %2 + %4 = sext i8 %3 to i32 + ret i32 %4 } -define i32 @load_sext_i16_to_i32(i16* %a) { -; CHECK-LABEL: load_sext_i16_to_i32 -; CHECK: ldrsh w0, [x0] +define i32 @load_unscaled_sext_i16_to_i32(i64 %a) { +; CHECK-LABEL: load_unscaled_sext_i16_to_i32 +; CHECK: ldursh w0, [x0, #-8] ; CHECK-NOT: sxth - %1 = load i16* %a - %2 = sext i16 %1 to i32 - ret i32 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i16* + %3 = load i16* %2 + %4 = sext i16 %3 to i32 + ret i32 %4 } -define i64 @load_sext_i8_to_i64(i8* %a) { -; CHECK-LABEL: load_sext_i8_to_i64 -; CHECK: ldrsb w0, [x0] +define i64 @load_unscaled_sext_i8_to_i64(i64 %a) { +; CHECK-LABEL: load_unscaled_sext_i8_to_i64 +; CHECK: ldursb x0, [x0, #-8] ; CHECK-NOT: sxtb - %1 = load i8* %a - %2 = sext i8 %1 to i64 - ret i64 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i8* + %3 = load i8* %2 + %4 = sext i8 %3 to i64 + ret i64 %4 } -define i64 @load_sext_i16_to_i64(i16* %a) { -; CHECK-LABEL: load_sext_i16_to_i64 -; CHECK: ldrsh w0, [x0] +define i64 @load_unscaled_sext_i16_to_i64(i64 %a) { +; CHECK-LABEL: load_unscaled_sext_i16_to_i64 +; CHECK: ldursh x0, [x0, #-8] ; CHECK-NOT: sxth - %1 = load i16* %a - %2 = sext i16 %1 to i64 - ret i64 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i16* + %3 = load i16* %2 + %4 = sext i16 %3 to i64 + ret i64 %4 } -define i64 @load_sext_i32_to_i64(i32* %a) { -; CHECK-LABEL: load_sext_i32_to_i64 -; CHECK: ldrsw x0, [x0] +define i64 @load_unscaled_sext_i32_to_i64(i64 %a) { +; CHECK-LABEL: load_unscaled_sext_i32_to_i64 +; CHECK: ldursw x0, [x0, #-8] ; CHECK-NOT: sxtw - %1 = load i32* %a - %2 = sext i32 %1 to i64 - ret i64 %2 + %1 = add i64 %a, -8 + %2 = inttoptr i64 %1 to i32* + %3 = load i32* %2 + %4 = sext i32 %3 to i64 + ret i64 %4 +} + +; Register +define i32 @load_register_zext_i8_to_i32(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_zext_i8_to_i32 +; CHECK: ldrb w0, [x0, x1] +; CHECK-NOT: uxtb + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i8* + %3 = load i8* %2 + %4 = zext i8 %3 to i32 + ret i32 %4 +} + +define i32 @load_register_zext_i16_to_i32(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_zext_i16_to_i32 +; CHECK: ldrh w0, [x0, x1] +; CHECK-NOT: uxth + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i16* + %3 = load i16* %2 + %4 = zext i16 %3 to i32 + ret i32 %4 +} + +define i64 @load_register_zext_i8_to_i64(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_zext_i8_to_i64 +; CHECK: ldrb w0, [x0, x1] +; CHECK-NOT: uxtb + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i8* + %3 = load i8* %2 + %4 = zext i8 %3 to i64 + ret i64 %4 +} + +define i64 @load_register_zext_i16_to_i64(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_zext_i16_to_i64 +; CHECK: ldrh w0, [x0, x1] +; CHECK-NOT: uxth + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i16* + %3 = load i16* %2 + %4 = zext i16 %3 to i64 + ret i64 %4 +} + +define i64 @load_register_zext_i32_to_i64(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_zext_i32_to_i64 +; CHECK: ldr w0, [x0, x1] +; CHECK-NOT: uxtw + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i32* + %3 = load i32* %2 + %4 = zext i32 %3 to i64 + ret i64 %4 +} + +define i32 @load_register_sext_i8_to_i32(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_sext_i8_to_i32 +; CHECK: ldrsb w0, [x0, x1] +; CHECK-NOT: sxtb + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i8* + %3 = load i8* %2 + %4 = sext i8 %3 to i32 + ret i32 %4 +} + +define i32 @load_register_sext_i16_to_i32(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_sext_i16_to_i32 +; CHECK: ldrsh w0, [x0, x1] +; CHECK-NOT: sxth + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i16* + %3 = load i16* %2 + %4 = sext i16 %3 to i32 + ret i32 %4 +} + +define i64 @load_register_sext_i8_to_i64(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_sext_i8_to_i64 +; CHECK: ldrsb x0, [x0, x1] +; CHECK-NOT: sxtb + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i8* + %3 = load i8* %2 + %4 = sext i8 %3 to i64 + ret i64 %4 +} + +define i64 @load_register_sext_i16_to_i64(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_sext_i16_to_i64 +; CHECK: ldrsh x0, [x0, x1] +; CHECK-NOT: sxth + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i16* + %3 = load i16* %2 + %4 = sext i16 %3 to i64 + ret i64 %4 +} + +define i64 @load_register_sext_i32_to_i64(i64 %a, i64 %b) { +; CHECK-LABEL: load_register_sext_i32_to_i64 +; CHECK: ldrsw x0, [x0, x1] +; CHECK-NOT: sxtw + %1 = add i64 %a, %b + %2 = inttoptr i64 %1 to i32* + %3 = load i32* %2 + %4 = sext i32 %3 to i64 + ret i64 %4 +} + +; Extend +define i32 @load_extend_zext_i8_to_i32(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_zext_i8_to_i32 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrb w0, [x0, [[REG]]] +; CHECK-NOT: uxtb + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i8* + %4 = load i8* %3 + %5 = zext i8 %4 to i32 + ret i32 %5 +} + +define i32 @load_extend_zext_i16_to_i32(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_zext_i16_to_i32 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrh w0, [x0, [[REG]]] +; CHECK-NOT: uxth + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i16* + %4 = load i16* %3 + %5 = zext i16 %4 to i32 + ret i32 %5 +} + +define i64 @load_extend_zext_i8_to_i64(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_zext_i8_to_i64 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrb w0, [x0, [[REG]]] +; CHECK-NOT: uxtb + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i8* + %4 = load i8* %3 + %5 = zext i8 %4 to i64 + ret i64 %5 +} + +define i64 @load_extend_zext_i16_to_i64(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_zext_i16_to_i64 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrh w0, [x0, [[REG]]] +; CHECK-NOT: uxth + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i16* + %4 = load i16* %3 + %5 = zext i16 %4 to i64 + ret i64 %5 +} + +define i64 @load_extend_zext_i32_to_i64(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_zext_i32_to_i64 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldr w0, [x0, [[REG]]] +; CHECK-NOT: uxtw + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i32* + %4 = load i32* %3 + %5 = zext i32 %4 to i64 + ret i64 %5 +} + +define i32 @load_extend_sext_i8_to_i32(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_sext_i8_to_i32 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrsb w0, [x0, [[REG]]] +; CHECK-NOT: sxtb + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i8* + %4 = load i8* %3 + %5 = sext i8 %4 to i32 + ret i32 %5 +} + +define i32 @load_extend_sext_i16_to_i32(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_sext_i16_to_i32 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrsh w0, [x0, [[REG]]] +; CHECK-NOT: sxth + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i16* + %4 = load i16* %3 + %5 = sext i16 %4 to i32 + ret i32 %5 +} + +define i64 @load_extend_sext_i8_to_i64(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_sext_i8_to_i64 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrsb x0, [x0, [[REG]]] +; CHECK-NOT: sxtb + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i8* + %4 = load i8* %3 + %5 = sext i8 %4 to i64 + ret i64 %5 +} + +define i64 @load_extend_sext_i16_to_i64(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_sext_i16_to_i64 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrsh x0, [x0, [[REG]]] +; CHECK-NOT: sxth + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i16* + %4 = load i16* %3 + %5 = sext i16 %4 to i64 + ret i64 %5 +} + +define i64 @load_extend_sext_i32_to_i64(i64 %a, i32 %b) { +; CHECK-LABEL: load_extend_sext_i32_to_i64 +; CHECK: sxtw [[REG:x[0-9]+]], w1 +; CHECK-NEXT: ldrsw x0, [x0, [[REG]]] +; CHECK-NOT: sxtw + %1 = sext i32 %b to i64 + %2 = add i64 %a, %1 + %3 = inttoptr i64 %2 to i32* + %4 = load i32* %3 + %5 = sext i32 %4 to i64 + ret i64 %5 } |