diff options
| author | Simon Pilgrim <llvm-dev@redking.me.uk> | 2018-03-04 15:00:19 +0000 | 
|---|---|---|
| committer | Simon Pilgrim <llvm-dev@redking.me.uk> | 2018-03-04 15:00:19 +0000 | 
| commit | 8197b04b9ba5a0f0fc4c9f16426d2eb26745e612 (patch) | |
| tree | db9bd895b3d56cffc49f594b5e5a74b5a3b945a8 | |
| parent | 3a76bc29d79cf7e2b9d118ffe52dc62a3165c1ca (diff) | |
| download | bcm5719-llvm-8197b04b9ba5a0f0fc4c9f16426d2eb26745e612.tar.gz bcm5719-llvm-8197b04b9ba5a0f0fc4c9f16426d2eb26745e612.zip  | |
[X86][X87] Add X87 folded integer arithmetic tests
Add tests for FIADD/FISUB/FISUBR/FIMUL/FIDIV/FIDIVR
Shows we have more FILD stack usage than necessary (arg load, spill, reload to x87)
llvm-svn: 326674
| -rw-r--r-- | llvm/test/CodeGen/X86/fp-arith.ll | 621 | 
1 files changed, 621 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/X86/fp-arith.ll b/llvm/test/CodeGen/X86/fp-arith.ll new file mode 100644 index 00000000000..e3711e04742 --- /dev/null +++ b/llvm/test/CodeGen/X86/fp-arith.ll @@ -0,0 +1,621 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=i386-unknown | FileCheck %s --check-prefixes=X86 +; RUN: llc < %s -mtriple=x86_64-unknown | FileCheck %s --check-prefixes=X64 + +; +; FADD +; + +define x86_fp80 @fiadd_fp80_i16(x86_fp80 %a0, i16 %a1) { +; X86-LABEL: fiadd_fp80_i16: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movzwl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fiadds {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fiadd_fp80_i16: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl %di, %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fiaddl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i16 %a1 to x86_fp80 +  %2 = fadd x86_fp80 %a0, %1 +  ret x86_fp80 %2 +} + +define x86_fp80 @fiadd_fp80_i16_ld(x86_fp80 %a0, i16 *%a1) { +; X86-LABEL: fiadd_fp80_i16_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movzwl (%eax), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fiadds {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fiadd_fp80_i16_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fiaddl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i16, i16 *%a1 +  %2 = sitofp i16 %1 to x86_fp80 +  %3 = fadd x86_fp80 %a0, %2 +  ret x86_fp80 %3 +} + +define x86_fp80 @fiadd_fp80_i32(x86_fp80 %a0, i32 %a1) { +; X86-LABEL: fiadd_fp80_i32: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fiaddl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fiadd_fp80_i32: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl %edi, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fiaddl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i32 %a1 to x86_fp80 +  %2 = fadd x86_fp80 %a0, %1 +  ret x86_fp80 %2 +} + +define x86_fp80 @fiadd_fp80_i32_ld(x86_fp80 %a0, i32 *%a1) { +; X86-LABEL: fiadd_fp80_i32_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl (%eax), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fiaddl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fiadd_fp80_i32_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fiaddl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i32, i32 *%a1 +  %2 = sitofp i32 %1 to x86_fp80 +  %3 = fadd x86_fp80 %a0, %2 +  ret x86_fp80 %3 +} + +; +; FSUB +; + +define x86_fp80 @fisub_fp80_i16(x86_fp80 %a0, i16 %a1) { +; X86-LABEL: fisub_fp80_i16: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movzwl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fisubs {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fisub_fp80_i16: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl %di, %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fisubl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i16 %a1 to x86_fp80 +  %2 = fsub x86_fp80 %a0, %1 +  ret x86_fp80 %2 +} + +define x86_fp80 @fisub_fp80_i16_ld(x86_fp80 %a0, i16 *%a1) { +; X86-LABEL: fisub_fp80_i16_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movzwl (%eax), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fisubs {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fisub_fp80_i16_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fisubl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i16, i16 *%a1 +  %2 = sitofp i16 %1 to x86_fp80 +  %3 = fsub x86_fp80 %a0, %2 +  ret x86_fp80 %3 +} + +define x86_fp80 @fisub_fp80_i32(x86_fp80 %a0, i32 %a1) { +; X86-LABEL: fisub_fp80_i32: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fisubl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fisub_fp80_i32: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl %edi, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fisubl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i32 %a1 to x86_fp80 +  %2 = fsub x86_fp80 %a0, %1 +  ret x86_fp80 %2 +} + +define x86_fp80 @fisub_fp80_i32_ld(x86_fp80 %a0, i32 *%a1) { +; X86-LABEL: fisub_fp80_i32_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl (%eax), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fisubl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fisub_fp80_i32_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fisubl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i32, i32 *%a1 +  %2 = sitofp i32 %1 to x86_fp80 +  %3 = fsub x86_fp80 %a0, %2 +  ret x86_fp80 %3 +} + +; +; FSUBR +; + +define x86_fp80 @fisubr_fp80_i16(x86_fp80 %a0, i16 %a1) { +; X86-LABEL: fisubr_fp80_i16: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movzwl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fisubrs {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fisubr_fp80_i16: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl %di, %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fisubrl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i16 %a1 to x86_fp80 +  %2 = fsub x86_fp80 %1, %a0 +  ret x86_fp80 %2 +} + +define x86_fp80 @fisubr_fp80_i16_ld(x86_fp80 %a0, i16 *%a1) { +; X86-LABEL: fisubr_fp80_i16_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movzwl (%eax), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fisubrs {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fisubr_fp80_i16_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fisubrl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i16, i16 *%a1 +  %2 = sitofp i16 %1 to x86_fp80 +  %3 = fsub x86_fp80 %2, %a0 +  ret x86_fp80 %3 +} + +define x86_fp80 @fisubr_fp80_i32(x86_fp80 %a0, i32 %a1) { +; X86-LABEL: fisubr_fp80_i32: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fisubrl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fisubr_fp80_i32: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl %edi, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fisubrl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i32 %a1 to x86_fp80 +  %2 = fsub x86_fp80 %1, %a0 +  ret x86_fp80 %2 +} + +define x86_fp80 @fisubr_fp80_i32_ld(x86_fp80 %a0, i32 *%a1) { +; X86-LABEL: fisubr_fp80_i32_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl (%eax), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fisubrl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fisubr_fp80_i32_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fisubrl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i32, i32 *%a1 +  %2 = sitofp i32 %1 to x86_fp80 +  %3 = fsub x86_fp80 %2, %a0 +  ret x86_fp80 %3 +} + +; +; FMUL +; + +define x86_fp80 @fimul_fp80_i16(x86_fp80 %a0, i16 %a1) { +; X86-LABEL: fimul_fp80_i16: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movzwl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fimuls {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fimul_fp80_i16: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl %di, %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fimull -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i16 %a1 to x86_fp80 +  %2 = fmul x86_fp80 %a0, %1 +  ret x86_fp80 %2 +} + +define x86_fp80 @fimul_fp80_i16_ld(x86_fp80 %a0, i16 *%a1) { +; X86-LABEL: fimul_fp80_i16_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movzwl (%eax), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fimuls {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fimul_fp80_i16_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fimull -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i16, i16 *%a1 +  %2 = sitofp i16 %1 to x86_fp80 +  %3 = fmul x86_fp80 %a0, %2 +  ret x86_fp80 %3 +} + +define x86_fp80 @fimul_fp80_i32(x86_fp80 %a0, i32 %a1) { +; X86-LABEL: fimul_fp80_i32: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fimull (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fimul_fp80_i32: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl %edi, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fimull -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i32 %a1 to x86_fp80 +  %2 = fmul x86_fp80 %a0, %1 +  ret x86_fp80 %2 +} + +define x86_fp80 @fimul_fp80_i32_ld(x86_fp80 %a0, i32 *%a1) { +; X86-LABEL: fimul_fp80_i32_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl (%eax), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fimull (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fimul_fp80_i32_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fimull -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i32, i32 *%a1 +  %2 = sitofp i32 %1 to x86_fp80 +  %3 = fmul x86_fp80 %a0, %2 +  ret x86_fp80 %3 +} + +; +; FDIV +; + +define x86_fp80 @fidiv_fp80_i16(x86_fp80 %a0, i16 %a1) { +; X86-LABEL: fidiv_fp80_i16: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movzwl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fidivs {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fidiv_fp80_i16: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl %di, %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fidivl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i16 %a1 to x86_fp80 +  %2 = fdiv x86_fp80 %a0, %1 +  ret x86_fp80 %2 +} + +define x86_fp80 @fidiv_fp80_i16_ld(x86_fp80 %a0, i16 *%a1) { +; X86-LABEL: fidiv_fp80_i16_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movzwl (%eax), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fidivs {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fidiv_fp80_i16_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fidivl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i16, i16 *%a1 +  %2 = sitofp i16 %1 to x86_fp80 +  %3 = fdiv x86_fp80 %a0, %2 +  ret x86_fp80 %3 +} + +define x86_fp80 @fidiv_fp80_i32(x86_fp80 %a0, i32 %a1) { +; X86-LABEL: fidiv_fp80_i32: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fidivl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fidiv_fp80_i32: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl %edi, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fidivl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i32 %a1 to x86_fp80 +  %2 = fdiv x86_fp80 %a0, %1 +  ret x86_fp80 %2 +} + +define x86_fp80 @fidiv_fp80_i32_ld(x86_fp80 %a0, i32 *%a1) { +; X86-LABEL: fidiv_fp80_i32_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl (%eax), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fidivl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fidiv_fp80_i32_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fidivl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i32, i32 *%a1 +  %2 = sitofp i32 %1 to x86_fp80 +  %3 = fdiv x86_fp80 %a0, %2 +  ret x86_fp80 %3 +} + +; +; FDIVR +; + +define x86_fp80 @fidivr_fp80_i16(x86_fp80 %a0, i16 %a1) { +; X86-LABEL: fidivr_fp80_i16: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movzwl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fidivrs {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fidivr_fp80_i16: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl %di, %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fidivrl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i16 %a1 to x86_fp80 +  %2 = fdiv x86_fp80 %1, %a0 +  ret x86_fp80 %2 +} + +define x86_fp80 @fidivr_fp80_i16_ld(x86_fp80 %a0, i16 *%a1) { +; X86-LABEL: fidivr_fp80_i16_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movzwl (%eax), %eax +; X86-NEXT:    movw %ax, {{[0-9]+}}(%esp) +; X86-NEXT:    fidivrs {{[0-9]+}}(%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fidivr_fp80_i16_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movswl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fidivrl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i16, i16 *%a1 +  %2 = sitofp i16 %1 to x86_fp80 +  %3 = fdiv x86_fp80 %2, %a0 +  ret x86_fp80 %3 +} + +define x86_fp80 @fidivr_fp80_i32(x86_fp80 %a0, i32 %a1) { +; X86-LABEL: fidivr_fp80_i32: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fidivrl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fidivr_fp80_i32: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl %edi, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fidivrl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = sitofp i32 %a1 to x86_fp80 +  %2 = fdiv x86_fp80 %1, %a0 +  ret x86_fp80 %2 +} + +define x86_fp80 @fidivr_fp80_i32_ld(x86_fp80 %a0, i32 *%a1) { +; X86-LABEL: fidivr_fp80_i32_ld: +; X86:       # %bb.0: +; X86-NEXT:    pushl %eax +; X86-NEXT:    .cfi_def_cfa_offset 8 +; X86-NEXT:    fldt {{[0-9]+}}(%esp) +; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax +; X86-NEXT:    movl (%eax), %eax +; X86-NEXT:    movl %eax, (%esp) +; X86-NEXT:    fidivrl (%esp) +; X86-NEXT:    popl %eax +; X86-NEXT:    retl +; +; X64-LABEL: fidivr_fp80_i32_ld: +; X64:       # %bb.0: +; X64-NEXT:    fldt {{[0-9]+}}(%rsp) +; X64-NEXT:    movl (%rdi), %eax +; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp) +; X64-NEXT:    fidivrl -{{[0-9]+}}(%rsp) +; X64-NEXT:    retq +  %1 = load i32, i32 *%a1 +  %2 = sitofp i32 %1 to x86_fp80 +  %3 = fdiv x86_fp80 %2, %a0 +  ret x86_fp80 %3 +}  | 

