summaryrefslogtreecommitdiffstats
path: root/llvm/test/CodeGen/RISCV/fp128.ll
diff options
context:
space:
mode:
Diffstat (limited to 'llvm/test/CodeGen/RISCV/fp128.ll')
-rw-r--r--llvm/test/CodeGen/RISCV/fp128.ll134
1 files changed, 134 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/RISCV/fp128.ll b/llvm/test/CodeGen/RISCV/fp128.ll
new file mode 100644
index 00000000000..32e559b75b5
--- /dev/null
+++ b/llvm/test/CodeGen/RISCV/fp128.ll
@@ -0,0 +1,134 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc -mtriple=riscv32 -verify-machineinstrs < %s \
+; RUN: | FileCheck -check-prefix=RV32I %s
+
+@x = local_unnamed_addr global fp128 0xL00000000000000007FFF000000000000, align 16
+@y = local_unnamed_addr global fp128 0xL00000000000000007FFF000000000000, align 16
+
+; Besides anything else, these tests help verify that libcall ABI lowering
+; works correctly
+
+define i32 @test_load_and_cmp() nounwind {
+; RV32I-LABEL: test_load_and_cmp:
+; RV32I: # %bb.0:
+; RV32I-NEXT: addi sp, sp, -48
+; RV32I-NEXT: sw ra, 44(sp)
+; RV32I-NEXT: sw s0, 40(sp)
+; RV32I-NEXT: addi s0, sp, 48
+; RV32I-NEXT: lui a0, %hi(y+12)
+; RV32I-NEXT: addi a0, a0, %lo(y+12)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -28(s0)
+; RV32I-NEXT: lui a0, %hi(y+8)
+; RV32I-NEXT: addi a0, a0, %lo(y+8)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -32(s0)
+; RV32I-NEXT: lui a0, %hi(y+4)
+; RV32I-NEXT: addi a0, a0, %lo(y+4)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -36(s0)
+; RV32I-NEXT: lui a0, %hi(y)
+; RV32I-NEXT: addi a0, a0, %lo(y)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -40(s0)
+; RV32I-NEXT: lui a0, %hi(x+12)
+; RV32I-NEXT: addi a0, a0, %lo(x+12)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -12(s0)
+; RV32I-NEXT: lui a0, %hi(x+8)
+; RV32I-NEXT: addi a0, a0, %lo(x+8)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -16(s0)
+; RV32I-NEXT: lui a0, %hi(x+4)
+; RV32I-NEXT: addi a0, a0, %lo(x+4)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -20(s0)
+; RV32I-NEXT: lui a0, %hi(x)
+; RV32I-NEXT: addi a0, a0, %lo(x)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -24(s0)
+; RV32I-NEXT: lui a0, %hi(__netf2)
+; RV32I-NEXT: addi a2, a0, %lo(__netf2)
+; RV32I-NEXT: addi a0, s0, -24
+; RV32I-NEXT: addi a1, s0, -40
+; RV32I-NEXT: jalr ra, a2, 0
+; RV32I-NEXT: xor a0, a0, zero
+; RV32I-NEXT: sltu a0, zero, a0
+; RV32I-NEXT: lw s0, 40(sp)
+; RV32I-NEXT: lw ra, 44(sp)
+; RV32I-NEXT: addi sp, sp, 48
+; RV32I-NEXT: jalr zero, ra, 0
+ %1 = load fp128, fp128* @x, align 16
+ %2 = load fp128, fp128* @y, align 16
+ %cmp = fcmp une fp128 %1, %2
+ %3 = zext i1 %cmp to i32
+ ret i32 %3
+}
+
+define i32 @test_add_and_fptosi() nounwind {
+; RV32I-LABEL: test_add_and_fptosi:
+; RV32I: # %bb.0:
+; RV32I-NEXT: addi sp, sp, -80
+; RV32I-NEXT: sw ra, 76(sp)
+; RV32I-NEXT: sw s0, 72(sp)
+; RV32I-NEXT: addi s0, sp, 80
+; RV32I-NEXT: lui a0, %hi(y+12)
+; RV32I-NEXT: addi a0, a0, %lo(y+12)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -44(s0)
+; RV32I-NEXT: lui a0, %hi(y+8)
+; RV32I-NEXT: addi a0, a0, %lo(y+8)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -48(s0)
+; RV32I-NEXT: lui a0, %hi(y+4)
+; RV32I-NEXT: addi a0, a0, %lo(y+4)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -52(s0)
+; RV32I-NEXT: lui a0, %hi(y)
+; RV32I-NEXT: addi a0, a0, %lo(y)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -56(s0)
+; RV32I-NEXT: lui a0, %hi(x+12)
+; RV32I-NEXT: addi a0, a0, %lo(x+12)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -28(s0)
+; RV32I-NEXT: lui a0, %hi(x+8)
+; RV32I-NEXT: addi a0, a0, %lo(x+8)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -32(s0)
+; RV32I-NEXT: lui a0, %hi(x+4)
+; RV32I-NEXT: addi a0, a0, %lo(x+4)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -36(s0)
+; RV32I-NEXT: lui a0, %hi(x)
+; RV32I-NEXT: addi a0, a0, %lo(x)
+; RV32I-NEXT: lw a0, 0(a0)
+; RV32I-NEXT: sw a0, -40(s0)
+; RV32I-NEXT: lui a0, %hi(__addtf3)
+; RV32I-NEXT: addi a3, a0, %lo(__addtf3)
+; RV32I-NEXT: addi a0, s0, -24
+; RV32I-NEXT: addi a1, s0, -40
+; RV32I-NEXT: addi a2, s0, -56
+; RV32I-NEXT: jalr ra, a3, 0
+; RV32I-NEXT: lw a0, -12(s0)
+; RV32I-NEXT: sw a0, -60(s0)
+; RV32I-NEXT: lw a0, -16(s0)
+; RV32I-NEXT: sw a0, -64(s0)
+; RV32I-NEXT: lw a0, -20(s0)
+; RV32I-NEXT: sw a0, -68(s0)
+; RV32I-NEXT: lw a0, -24(s0)
+; RV32I-NEXT: sw a0, -72(s0)
+; RV32I-NEXT: lui a0, %hi(__fixtfsi)
+; RV32I-NEXT: addi a1, a0, %lo(__fixtfsi)
+; RV32I-NEXT: addi a0, s0, -72
+; RV32I-NEXT: jalr ra, a1, 0
+; RV32I-NEXT: lw s0, 72(sp)
+; RV32I-NEXT: lw ra, 76(sp)
+; RV32I-NEXT: addi sp, sp, 80
+; RV32I-NEXT: jalr zero, ra, 0
+ %1 = load fp128, fp128* @x, align 16
+ %2 = load fp128, fp128* @y, align 16
+ %3 = fadd fp128 %1, %2
+ %4 = fptosi fp128 %3 to i32
+ ret i32 %4
+}
OpenPOWER on IntegriCloud