diff options
Diffstat (limited to 'llvm/test/CodeGen/X86/urem-seteq-optsize.ll')
-rw-r--r-- | llvm/test/CodeGen/X86/urem-seteq-optsize.ll | 81 |
1 files changed, 81 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/X86/urem-seteq-optsize.ll b/llvm/test/CodeGen/X86/urem-seteq-optsize.ll new file mode 100644 index 00000000000..1d79aadff6a --- /dev/null +++ b/llvm/test/CodeGen/X86/urem-seteq-optsize.ll @@ -0,0 +1,81 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc -mtriple=i686-unknown-linux-gnu < %s | FileCheck %s --check-prefixes=CHECK,X86 +; RUN: llc -mtriple=x86_64-unknown-linux-gnu < %s | FileCheck %s --check-prefixes=CHECK,X64,NOBMI2 +; RUN: llc -mtriple=x86_64-unknown-linux-gnu -mattr=+bmi2 < %s | FileCheck %s --check-prefixes=CHECK,X64,BMI2 + +; On X86, division in expensive. BuildRemEqFold should therefore run even +; when optimizing for size. Only optimizing for minimum size retains a plain div. + +define i32 @test_minsize(i32 %X) optsize minsize nounwind readnone { +; X86-LABEL: test_minsize: +; X86: # %bb.0: +; X86-NEXT: movl {{[0-9]+}}(%esp), %eax +; X86-NEXT: pushl $5 +; X86-NEXT: popl %ecx +; X86-NEXT: xorl %edx, %edx +; X86-NEXT: divl %ecx +; X86-NEXT: testl %edx, %edx +; X86-NEXT: je .LBB0_1 +; X86-NEXT: # %bb.2: +; X86-NEXT: pushl $-10 +; X86-NEXT: popl %eax +; X86-NEXT: retl +; X86-NEXT: .LBB0_1: +; X86-NEXT: pushl $42 +; X86-NEXT: popl %eax +; X86-NEXT: retl +; +; X64-LABEL: test_minsize: +; X64: # %bb.0: +; X64-NEXT: pushq $5 +; X64-NEXT: popq %rcx +; X64-NEXT: xorl %edx, %edx +; X64-NEXT: movl %edi, %eax +; X64-NEXT: divl %ecx +; X64-NEXT: testl %edx, %edx +; X64-NEXT: pushq $42 +; X64-NEXT: popq %rcx +; X64-NEXT: pushq $-10 +; X64-NEXT: popq %rax +; X64-NEXT: cmovel %ecx, %eax +; X64-NEXT: retq + %rem = urem i32 %X, 5 + %cmp = icmp eq i32 %rem, 0 + %ret = select i1 %cmp, i32 42, i32 -10 + ret i32 %ret +} + +define i32 @test_optsize(i32 %X) optsize nounwind readnone { +; X86-LABEL: test_optsize: +; X86: # %bb.0: +; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx +; X86-NEXT: movl $-858993459, %edx # imm = 0xCCCCCCCD +; X86-NEXT: movl %ecx, %eax +; X86-NEXT: mull %edx +; X86-NEXT: shrl $2, %edx +; X86-NEXT: leal (%edx,%edx,4), %eax +; X86-NEXT: cmpl %eax, %ecx +; X86-NEXT: movl $42, %eax +; X86-NEXT: je .LBB1_2 +; X86-NEXT: # %bb.1: +; X86-NEXT: movl $-10, %eax +; X86-NEXT: .LBB1_2: +; X86-NEXT: retl +; +; X64-LABEL: test_optsize: +; X64: # %bb.0: +; X64-NEXT: movl %edi, %eax +; X64-NEXT: movl $3435973837, %ecx # imm = 0xCCCCCCCD +; X64-NEXT: imulq %rax, %rcx +; X64-NEXT: shrq $34, %rcx +; X64-NEXT: leal (%rcx,%rcx,4), %eax +; X64-NEXT: cmpl %eax, %edi +; X64-NEXT: movl $42, %ecx +; X64-NEXT: movl $-10, %eax +; X64-NEXT: cmovel %ecx, %eax +; X64-NEXT: retq + %rem = urem i32 %X, 5 + %cmp = icmp eq i32 %rem, 0 + %ret = select i1 %cmp, i32 42, i32 -10 + ret i32 %ret +} |