diff options
Diffstat (limited to 'llvm')
-rw-r--r-- | llvm/test/CodeGen/X86/uint64-to-float.ll | 57 |
1 files changed, 41 insertions, 16 deletions
diff --git a/llvm/test/CodeGen/X86/uint64-to-float.ll b/llvm/test/CodeGen/X86/uint64-to-float.ll index 8a8d29bfda5..ab9467033a1 100644 --- a/llvm/test/CodeGen/X86/uint64-to-float.ll +++ b/llvm/test/CodeGen/X86/uint64-to-float.ll @@ -1,23 +1,48 @@ -; RUN: llc < %s -mcpu=generic -march=x86-64 | FileCheck %s +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=i686-apple-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X86 +; RUN: llc < %s -mtriple=x86_64-apple-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X64 + ; Verify that we are using the efficient uitofp --> sitofp lowering illustrated ; by the compiler_rt implementation of __floatundisf. ; <rdar://problem/8493982> -target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64" -target triple = "x86_64-apple-darwin10.0.0" - -; CHECK: %entry -; CHECK-NEXT: testq %rdi, %rdi -; CHECK-NEXT: js LBB0_1 -; CHECK: cvtsi2ss -; CHECK-NEXT: ret -; CHECK: LBB0_1 -; CHECK: movq -; CHECK-NEXT: shrq -; CHECK-NEXT: andl -; CHECK-NEXT: orq -; CHECK-NEXT: cvtsi2ss -define float @test(i64 %a) { +define float @test(i64 %a) nounwind { +; X86-LABEL: test: +; X86: # BB#0: # %entry +; X86-NEXT: pushl %ebp +; X86-NEXT: movl %esp, %ebp +; X86-NEXT: andl $-8, %esp +; X86-NEXT: subl $16, %esp +; X86-NEXT: movq {{.*#+}} xmm0 = mem[0],zero +; X86-NEXT: movq %xmm0, {{[0-9]+}}(%esp) +; X86-NEXT: xorl %eax, %eax +; X86-NEXT: cmpl $0, 12(%ebp) +; X86-NEXT: setns %al +; X86-NEXT: fildll {{[0-9]+}}(%esp) +; X86-NEXT: fadds {{\.LCPI.*}}(,%eax,4) +; X86-NEXT: fstps {{[0-9]+}}(%esp) +; X86-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero +; X86-NEXT: movss %xmm0, (%esp) +; X86-NEXT: flds (%esp) +; X86-NEXT: movl %ebp, %esp +; X86-NEXT: popl %ebp +; X86-NEXT: retl +; +; X64-LABEL: test: +; X64: # BB#0: # %entry +; X64-NEXT: testq %rdi, %rdi +; X64-NEXT: js .LBB0_1 +; X64-NEXT: # BB#2: # %entry +; X64-NEXT: cvtsi2ssq %rdi, %xmm0 +; X64-NEXT: retq +; X64-NEXT: .LBB0_1: +; X64-NEXT: movq %rdi, %rax +; X64-NEXT: shrq %rax +; X64-NEXT: andl $1, %edi +; X64-NEXT: orq %rax, %rdi +; X64-NEXT: cvtsi2ssq %rdi, %xmm0 +; X64-NEXT: addss %xmm0, %xmm0 +; X64-NEXT: retq entry: %b = uitofp i64 %a to float ret float %b |