summaryrefslogtreecommitdiffstats
path: root/llvm/test/CodeGen/Mips/llvm-ir/sdiv.ll
diff options
context:
space:
mode:
Diffstat (limited to 'llvm/test/CodeGen/Mips/llvm-ir/sdiv.ll')
-rw-r--r--llvm/test/CodeGen/Mips/llvm-ir/sdiv.ll627
1 files changed, 471 insertions, 156 deletions
diff --git a/llvm/test/CodeGen/Mips/llvm-ir/sdiv.ll b/llvm/test/CodeGen/Mips/llvm-ir/sdiv.ll
index ee86f4b7baf..03b831191a8 100644
--- a/llvm/test/CodeGen/Mips/llvm-ir/sdiv.ll
+++ b/llvm/test/CodeGen/Mips/llvm-ir/sdiv.ll
@@ -1,192 +1,507 @@
-; RUN: llc < %s -march=mips -mcpu=mips2 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,NOT-R2-R6,GP32
-; RUN: llc < %s -march=mips -mcpu=mips32 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,NOT-R2-R6,GP32
-; RUN: llc < %s -march=mips -mcpu=mips32r2 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,R2-R5,GP32
-; RUN: llc < %s -march=mips -mcpu=mips32r3 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,R2-R5,GP32
-; RUN: llc < %s -march=mips -mcpu=mips32r5 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,R2-R5,GP32
-; RUN: llc < %s -march=mips -mcpu=mips32r6 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,R6,GP32
-
-; RUN: llc < %s -march=mips64 -mcpu=mips3 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,NOT-R2-R6,GP64-NOT-R6
-; RUN: llc < %s -march=mips64 -mcpu=mips4 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,NOT-R2-R6,GP64-NOT-R6
-; RUN: llc < %s -march=mips64 -mcpu=mips64 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,NOT-R2-R6,GP64-NOT-R6
-; RUN: llc < %s -march=mips64 -mcpu=mips64r2 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,R2-R5,GP64-NOT-R6
-; RUN: llc < %s -march=mips64 -mcpu=mips64r3 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,R2-R5,GP64-NOT-R6
-; RUN: llc < %s -march=mips64 -mcpu=mips64r5 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,NOT-R6,R2-R5,GP64-NOT-R6
-; RUN: llc < %s -march=mips64 -mcpu=mips64r6 -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,R6,64R6
-
-; RUN: llc < %s -march=mips -mcpu=mips32r3 -mattr=+micromips -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,MMR3,MM32
-; RUN: llc < %s -march=mips -mcpu=mips32r6 -mattr=+micromips -relocation-model=pic | FileCheck %s \
-; RUN: -check-prefixes=ALL,MMR6,MM32
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -mtriple=mips -mcpu=mips2 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP32,GP32R0R2
+; RUN: llc < %s -mtriple=mips -mcpu=mips32 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP32,GP32R0R2
+; RUN: llc < %s -mtriple=mips -mcpu=mips32r2 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP32,GP32R2R5
+; RUN: llc < %s -mtriple=mips -mcpu=mips32r3 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP32,GP32R2R5
+; RUN: llc < %s -mtriple=mips -mcpu=mips32r5 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP32,GP32R2R5
+; RUN: llc < %s -mtriple=mips -mcpu=mips32r6 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefix=GP32R6
+
+; RUN: llc < %s -mtriple=mips64 -mcpu=mips3 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP64,GP64R0R1
+; RUN: llc < %s -mtriple=mips64 -mcpu=mips4 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP64,GP64R0R1
+; RUN: llc < %s -mtriple=mips64 -mcpu=mips64 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP64,GP64R0R1
+; RUN: llc < %s -mtriple=mips64 -mcpu=mips64r2 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP64,GP64R2R5
+; RUN: llc < %s -mtriple=mips64 -mcpu=mips64r3 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP64,GP64R2R5
+; RUN: llc < %s -mtriple=mips64 -mcpu=mips64r5 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefixes=GP64,GP64R2R5
+; RUN: llc < %s -mtriple=mips64 -mcpu=mips64r6 -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefix=GP64R6
+
+; RUN: llc < %s -mtriple=mips -mcpu=mips32r3 -mattr=+micromips -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefix=MMR3
+; RUN: llc < %s -mtriple=mips -mcpu=mips32r6 -mattr=+micromips -relocation-model=pic \
+; RUN: | FileCheck %s -check-prefix=MMR6
define signext i1 @sdiv_i1(i1 signext %a, i1 signext %b) {
+; GP32-LABEL: sdiv_i1:
+; GP32: # %bb.0: # %entry
+; GP32-NEXT: div $zero, $4, $5
+; GP32-NEXT: teq $5, $zero, 7
+; GP32-NEXT: mflo $1
+; GP32-NEXT: andi $1, $1, 1
+; GP32-NEXT: jr $ra
+; GP32-NEXT: negu $2, $1
+;
+; GP32R6-LABEL: sdiv_i1:
+; GP32R6: # %bb.0: # %entry
+; GP32R6-NEXT: div $1, $4, $5
+; GP32R6-NEXT: teq $5, $zero, 7
+; GP32R6-NEXT: andi $1, $1, 1
+; GP32R6-NEXT: jr $ra
+; GP32R6-NEXT: negu $2, $1
+;
+; GP64-LABEL: sdiv_i1:
+; GP64: # %bb.0: # %entry
+; GP64-NEXT: div $zero, $4, $5
+; GP64-NEXT: teq $5, $zero, 7
+; GP64-NEXT: mflo $1
+; GP64-NEXT: andi $1, $1, 1
+; GP64-NEXT: jr $ra
+; GP64-NEXT: negu $2, $1
+;
+; GP64R6-LABEL: sdiv_i1:
+; GP64R6: # %bb.0: # %entry
+; GP64R6-NEXT: div $1, $4, $5
+; GP64R6-NEXT: teq $5, $zero, 7
+; GP64R6-NEXT: andi $1, $1, 1
+; GP64R6-NEXT: jr $ra
+; GP64R6-NEXT: negu $2, $1
+;
+; MMR3-LABEL: sdiv_i1:
+; MMR3: # %bb.0: # %entry
+; MMR3-NEXT: div $zero, $4, $5
+; MMR3-NEXT: teq $5, $zero, 7
+; MMR3-NEXT: mflo16 $2
+; MMR3-NEXT: andi16 $2, $2, 1
+; MMR3-NEXT: li16 $3, 0
+; MMR3-NEXT: subu16 $2, $3, $2
+; MMR3-NEXT: jrc $ra
+;
+; MMR6-LABEL: sdiv_i1:
+; MMR6: # %bb.0: # %entry
+; MMR6-NEXT: div $2, $4, $5
+; MMR6-NEXT: teq $5, $zero, 7
+; MMR6-NEXT: andi16 $2, $2, 1
+; MMR6-NEXT: li16 $3, 0
+; MMR6-NEXT: subu16 $2, $3, $2
+; MMR6-NEXT: jrc $ra
entry:
-; ALL-LABEL: sdiv_i1:
-
- ; NOT-R6: div $zero, $4, $5
- ; NOT-R6: teq $5, $zero, 7
- ; NOT-R6: mflo $[[T0:[0-9]+]]
- ; FIXME: The andi/negu instructions are redundant since div is signed.
- ; NOT-R6: andi $[[T0]], $[[T0]], 1
- ; NOT-R6: negu $2, $[[T0]]
-
- ; R6: div $[[T0:[0-9]+]], $4, $5
- ; R6: teq $5, $zero, 7
- ; FIXME: The andi/negu instructions are redundant since div is signed.
- ; R6: andi $[[T0]], $[[T0]], 1
- ; R6: negu $2, $[[T0]]
-
- ; MMR3: div $zero, $4, $5
- ; MMR3: teq $5, $zero, 7
- ; MMR3: mflo16 $[[T0:[0-9]+]]
- ; MMR3: andi16 $[[T0]], $[[T0]], 1
- ; MMR3: li16 $[[T1:[0-9]+]], 0
- ; MMR3: subu16 $2, $[[T1]], $[[T0]]
-
- ; MMR6: div $[[T0:[0-9]+]], $4, $5
- ; MMR6: teq $5, $zero, 7
- ; MMR6: andi16 $[[T0]], $[[T0]], 1
- ; MMR6: li16 $[[T1:[0-9]+]], 0
- ; MMR6: subu16 $2, $[[T1]], $[[T0]]
-
%r = sdiv i1 %a, %b
ret i1 %r
}
define signext i8 @sdiv_i8(i8 signext %a, i8 signext %b) {
+; GP32R0R2-LABEL: sdiv_i8:
+; GP32R0R2: # %bb.0: # %entry
+; GP32R0R2-NEXT: div $zero, $4, $5
+; GP32R0R2-NEXT: teq $5, $zero, 7
+; GP32R0R2-NEXT: mflo $1
+; GP32R0R2-NEXT: sll $1, $1, 24
+; GP32R0R2-NEXT: jr $ra
+; GP32R0R2-NEXT: sra $2, $1, 24
+;
+; GP32R2R5-LABEL: sdiv_i8:
+; GP32R2R5: # %bb.0: # %entry
+; GP32R2R5-NEXT: div $zero, $4, $5
+; GP32R2R5-NEXT: teq $5, $zero, 7
+; GP32R2R5-NEXT: mflo $1
+; GP32R2R5-NEXT: jr $ra
+; GP32R2R5-NEXT: seb $2, $1
+;
+; GP32R6-LABEL: sdiv_i8:
+; GP32R6: # %bb.0: # %entry
+; GP32R6-NEXT: div $1, $4, $5
+; GP32R6-NEXT: teq $5, $zero, 7
+; GP32R6-NEXT: jr $ra
+; GP32R6-NEXT: seb $2, $1
+;
+; GP64R0R1-LABEL: sdiv_i8:
+; GP64R0R1: # %bb.0: # %entry
+; GP64R0R1-NEXT: div $zero, $4, $5
+; GP64R0R1-NEXT: teq $5, $zero, 7
+; GP64R0R1-NEXT: mflo $1
+; GP64R0R1-NEXT: sll $1, $1, 24
+; GP64R0R1-NEXT: jr $ra
+; GP64R0R1-NEXT: sra $2, $1, 24
+;
+; GP64R2R5-LABEL: sdiv_i8:
+; GP64R2R5: # %bb.0: # %entry
+; GP64R2R5-NEXT: div $zero, $4, $5
+; GP64R2R5-NEXT: teq $5, $zero, 7
+; GP64R2R5-NEXT: mflo $1
+; GP64R2R5-NEXT: jr $ra
+; GP64R2R5-NEXT: seb $2, $1
+;
+; GP64R6-LABEL: sdiv_i8:
+; GP64R6: # %bb.0: # %entry
+; GP64R6-NEXT: div $1, $4, $5
+; GP64R6-NEXT: teq $5, $zero, 7
+; GP64R6-NEXT: jr $ra
+; GP64R6-NEXT: seb $2, $1
+;
+; MMR3-LABEL: sdiv_i8:
+; MMR3: # %bb.0: # %entry
+; MMR3-NEXT: div $zero, $4, $5
+; MMR3-NEXT: teq $5, $zero, 7
+; MMR3-NEXT: mflo16 $1
+; MMR3-NEXT: jr $ra
+; MMR3-NEXT: seb $2, $1
+;
+; MMR6-LABEL: sdiv_i8:
+; MMR6: # %bb.0: # %entry
+; MMR6-NEXT: div $1, $4, $5
+; MMR6-NEXT: teq $5, $zero, 7
+; MMR6-NEXT: seb $2, $1
+; MMR6-NEXT: jrc $ra
entry:
-; ALL-LABEL: sdiv_i8:
-
- ; NOT-R2-R6: div $zero, $4, $5
- ; NOT-R2-R6: teq $5, $zero, 7
- ; NOT-R2-R6: mflo $[[T0:[0-9]+]]
- ; FIXME: The sll/sra instructions are redundant since div is signed.
- ; NOT-R2-R6: sll $[[T1:[0-9]+]], $[[T0]], 24
- ; NOT-R2-R6: sra $2, $[[T1]], 24
-
- ; R2-R5: div $zero, $4, $5
- ; R2-R5: teq $5, $zero, 7
- ; R2-R5: mflo $[[T0:[0-9]+]]
- ; FIXME: This instruction is redundant.
- ; R2-R5: seb $2, $[[T0]]
-
- ; R6: div $[[T0:[0-9]+]], $4, $5
- ; R6: teq $5, $zero, 7
- ; FIXME: This instruction is redundant.
- ; R6: seb $2, $[[T0]]
-
- ; MMR3: div $zero, $4, $5
- ; MMR3: teq $5, $zero, 7
- ; MMR3: mflo16 $[[T0:[0-9]+]]
- ; MMR3: seb $2, $[[T0]]
-
- ; MMR6: div $[[T0:[0-9]+]], $4, $5
- ; MMR6: teq $5, $zero, 7
- ; MMR6: seb $2, $[[T0]]
-
%r = sdiv i8 %a, %b
ret i8 %r
}
define signext i16 @sdiv_i16(i16 signext %a, i16 signext %b) {
+; GP32R0R2-LABEL: sdiv_i16:
+; GP32R0R2: # %bb.0: # %entry
+; GP32R0R2-NEXT: div $zero, $4, $5
+; GP32R0R2-NEXT: teq $5, $zero, 7
+; GP32R0R2-NEXT: mflo $1
+; GP32R0R2-NEXT: sll $1, $1, 16
+; GP32R0R2-NEXT: jr $ra
+; GP32R0R2-NEXT: sra $2, $1, 16
+;
+; GP32R2R5-LABEL: sdiv_i16:
+; GP32R2R5: # %bb.0: # %entry
+; GP32R2R5-NEXT: div $zero, $4, $5
+; GP32R2R5-NEXT: teq $5, $zero, 7
+; GP32R2R5-NEXT: mflo $1
+; GP32R2R5-NEXT: jr $ra
+; GP32R2R5-NEXT: seh $2, $1
+;
+; GP32R6-LABEL: sdiv_i16:
+; GP32R6: # %bb.0: # %entry
+; GP32R6-NEXT: div $1, $4, $5
+; GP32R6-NEXT: teq $5, $zero, 7
+; GP32R6-NEXT: jr $ra
+; GP32R6-NEXT: seh $2, $1
+;
+; GP64R0R1-LABEL: sdiv_i16:
+; GP64R0R1: # %bb.0: # %entry
+; GP64R0R1-NEXT: div $zero, $4, $5
+; GP64R0R1-NEXT: teq $5, $zero, 7
+; GP64R0R1-NEXT: mflo $1
+; GP64R0R1-NEXT: sll $1, $1, 16
+; GP64R0R1-NEXT: jr $ra
+; GP64R0R1-NEXT: sra $2, $1, 16
+;
+; GP64R2R5-LABEL: sdiv_i16:
+; GP64R2R5: # %bb.0: # %entry
+; GP64R2R5-NEXT: div $zero, $4, $5
+; GP64R2R5-NEXT: teq $5, $zero, 7
+; GP64R2R5-NEXT: mflo $1
+; GP64R2R5-NEXT: jr $ra
+; GP64R2R5-NEXT: seh $2, $1
+;
+; GP64R6-LABEL: sdiv_i16:
+; GP64R6: # %bb.0: # %entry
+; GP64R6-NEXT: div $1, $4, $5
+; GP64R6-NEXT: teq $5, $zero, 7
+; GP64R6-NEXT: jr $ra
+; GP64R6-NEXT: seh $2, $1
+;
+; MMR3-LABEL: sdiv_i16:
+; MMR3: # %bb.0: # %entry
+; MMR3-NEXT: div $zero, $4, $5
+; MMR3-NEXT: teq $5, $zero, 7
+; MMR3-NEXT: mflo16 $1
+; MMR3-NEXT: jr $ra
+; MMR3-NEXT: seh $2, $1
+;
+; MMR6-LABEL: sdiv_i16:
+; MMR6: # %bb.0: # %entry
+; MMR6-NEXT: div $1, $4, $5
+; MMR6-NEXT: teq $5, $zero, 7
+; MMR6-NEXT: seh $2, $1
+; MMR6-NEXT: jrc $ra
entry:
-; ALL-LABEL: sdiv_i16:
-
- ; NOT-R2-R6: div $zero, $4, $5
- ; NOT-R2-R6: teq $5, $zero, 7
- ; NOT-R2-R6: mflo $[[T0:[0-9]+]]
- ; FIXME: The sll/sra instructions are redundant since div is signed.
- ; NOT-R2-R6: sll $[[T1:[0-9]+]], $[[T0]], 16
- ; NOT-R2-R6: sra $2, $[[T1]], 16
-
- ; R2-R5: div $zero, $4, $5
- ; R2-R5: teq $5, $zero, 7
- ; R2-R5: mflo $[[T0:[0-9]+]]
- ; FIXME: This is instruction is redundant since div is signed.
- ; R2-R5: seh $2, $[[T0]]
-
- ; R6: div $[[T0:[0-9]+]], $4, $5
- ; R6: teq $5, $zero, 7
- ; FIXME: This is instruction is redundant since div is signed.
- ; R6: seh $2, $[[T0]]
-
- ; MMR3: div $zero, $4, $5
- ; MMR3: teq $5, $zero, 7
- ; MMR3: mflo16 $[[T0:[0-9]+]]
- ; MMR3: seh $2, $[[T0]]
-
- ; MMR6: div $[[T0:[0-9]+]], $4, $5
- ; MMR6: teq $5, $zero, 7
- ; MMR6: seh $2, $[[T0]]
-
%r = sdiv i16 %a, %b
ret i16 %r
}
define signext i32 @sdiv_i32(i32 signext %a, i32 signext %b) {
+; GP32-LABEL: sdiv_i32:
+; GP32: # %bb.0: # %entry
+; GP32-NEXT: div $zero, $4, $5
+; GP32-NEXT: teq $5, $zero, 7
+; GP32-NEXT: jr $ra
+; GP32-NEXT: mflo $2
+;
+; GP32R6-LABEL: sdiv_i32:
+; GP32R6: # %bb.0: # %entry
+; GP32R6-NEXT: div $2, $4, $5
+; GP32R6-NEXT: teq $5, $zero, 7
+; GP32R6-NEXT: jrc $ra
+;
+; GP64-LABEL: sdiv_i32:
+; GP64: # %bb.0: # %entry
+; GP64-NEXT: div $zero, $4, $5
+; GP64-NEXT: teq $5, $zero, 7
+; GP64-NEXT: jr $ra
+; GP64-NEXT: mflo $2
+;
+; GP64R6-LABEL: sdiv_i32:
+; GP64R6: # %bb.0: # %entry
+; GP64R6-NEXT: div $2, $4, $5
+; GP64R6-NEXT: teq $5, $zero, 7
+; GP64R6-NEXT: jrc $ra
+;
+; MMR3-LABEL: sdiv_i32:
+; MMR3: # %bb.0: # %entry
+; MMR3-NEXT: div $zero, $4, $5
+; MMR3-NEXT: teq $5, $zero, 7
+; MMR3-NEXT: mflo16 $2
+; MMR3-NEXT: jrc $ra
+;
+; MMR6-LABEL: sdiv_i32:
+; MMR6: # %bb.0: # %entry
+; MMR6-NEXT: div $2, $4, $5
+; MMR6-NEXT: teq $5, $zero, 7
+; MMR6-NEXT: jrc $ra
entry:
-; ALL-LABEL: sdiv_i32:
-
- ; NOT-R6: div $zero, $4, $5
- ; NOT-R6: teq $5, $zero, 7
- ; NOT-R6: mflo $2
-
- ; R6: div $2, $4, $5
- ; R6: teq $5, $zero, 7
-
- ; MMR3: div $zero, $4, $5
- ; MMR3: teq $5, $zero, 7
- ; MMR3: mflo16 $2
-
- ; MMR6: div $2, $4, $5
- ; MMR6: teq $5, $zero, 7
-
%r = sdiv i32 %a, %b
ret i32 %r
}
define signext i64 @sdiv_i64(i64 signext %a, i64 signext %b) {
+; GP32-LABEL: sdiv_i64:
+; GP32: # %bb.0: # %entry
+; GP32-NEXT: lui $2, %hi(_gp_disp)
+; GP32-NEXT: addiu $2, $2, %lo(_gp_disp)
+; GP32-NEXT: addiu $sp, $sp, -24
+; GP32-NEXT: .cfi_def_cfa_offset 24
+; GP32-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
+; GP32-NEXT: .cfi_offset 31, -4
+; GP32-NEXT: addu $gp, $2, $25
+; GP32-NEXT: lw $25, %call16(__divdi3)($gp)
+; GP32-NEXT: jalr $25
+; GP32-NEXT: nop
+; GP32-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
+; GP32-NEXT: jr $ra
+; GP32-NEXT: addiu $sp, $sp, 24
+;
+; GP32R6-LABEL: sdiv_i64:
+; GP32R6: # %bb.0: # %entry
+; GP32R6-NEXT: lui $2, %hi(_gp_disp)
+; GP32R6-NEXT: addiu $2, $2, %lo(_gp_disp)
+; GP32R6-NEXT: addiu $sp, $sp, -24
+; GP32R6-NEXT: .cfi_def_cfa_offset 24
+; GP32R6-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
+; GP32R6-NEXT: .cfi_offset 31, -4
+; GP32R6-NEXT: addu $gp, $2, $25
+; GP32R6-NEXT: lw $25, %call16(__divdi3)($gp)
+; GP32R6-NEXT: jalrc $25
+; GP32R6-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
+; GP32R6-NEXT: jr $ra
+; GP32R6-NEXT: addiu $sp, $sp, 24
+;
+; GP64-LABEL: sdiv_i64:
+; GP64: # %bb.0: # %entry
+; GP64-NEXT: ddiv $zero, $4, $5
+; GP64-NEXT: teq $5, $zero, 7
+; GP64-NEXT: jr $ra
+; GP64-NEXT: mflo $2
+;
+; GP64R6-LABEL: sdiv_i64:
+; GP64R6: # %bb.0: # %entry
+; GP64R6-NEXT: ddiv $2, $4, $5
+; GP64R6-NEXT: teq $5, $zero, 7
+; GP64R6-NEXT: jrc $ra
+;
+; MMR3-LABEL: sdiv_i64:
+; MMR3: # %bb.0: # %entry
+; MMR3-NEXT: lui $2, %hi(_gp_disp)
+; MMR3-NEXT: addiu $2, $2, %lo(_gp_disp)
+; MMR3-NEXT: addiusp -24
+; MMR3-NEXT: .cfi_def_cfa_offset 24
+; MMR3-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
+; MMR3-NEXT: .cfi_offset 31, -4
+; MMR3-NEXT: addu $2, $2, $25
+; MMR3-NEXT: lw $25, %call16(__divdi3)($2)
+; MMR3-NEXT: move $gp, $2
+; MMR3-NEXT: jalr $25
+; MMR3-NEXT: nop
+; MMR3-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
+; MMR3-NEXT: addiusp 24
+; MMR3-NEXT: jrc $ra
+;
+; MMR6-LABEL: sdiv_i64:
+; MMR6: # %bb.0: # %entry
+; MMR6-NEXT: lui $2, %hi(_gp_disp)
+; MMR6-NEXT: addiu $2, $2, %lo(_gp_disp)
+; MMR6-NEXT: addiu $sp, $sp, -24
+; MMR6-NEXT: .cfi_def_cfa_offset 24
+; MMR6-NEXT: sw $ra, 20($sp) # 4-byte Folded Spill
+; MMR6-NEXT: .cfi_offset 31, -4
+; MMR6-NEXT: addu $2, $2, $25
+; MMR6-NEXT: lw $25, %call16(__divdi3)($2)
+; MMR6-NEXT: move $gp, $2
+; MMR6-NEXT: jalr $25
+; MMR6-NEXT: lw $ra, 20($sp) # 4-byte Folded Reload
+; MMR6-NEXT: addiu $sp, $sp, 24
+; MMR6-NEXT: jrc $ra
entry:
-; ALL-LABEL: sdiv_i64:
-
- ; GP32: lw $25, %call16(__divdi3)($gp)
-
- ; GP64-NOT-R6: ddiv $zero, $4, $5
- ; GP64-NOT-R6: teq $5, $zero, 7
- ; GP64-NOT-R6: mflo $2
-
- ; 64R6: ddiv $2, $4, $5
- ; 64R6: teq $5, $zero, 7
-
- ; MM32: lw $25, %call16(__divdi3)($2)
-
%r = sdiv i64 %a, %b
ret i64 %r
}
define signext i128 @sdiv_i128(i128 signext %a, i128 signext %b) {
+; GP32-LABEL: sdiv_i128:
+; GP32: # %bb.0: # %entry
+; GP32-NEXT: lui $2, %hi(_gp_disp)
+; GP32-NEXT: addiu $2, $2, %lo(_gp_disp)
+; GP32-NEXT: addiu $sp, $sp, -40
+; GP32-NEXT: .cfi_def_cfa_offset 40
+; GP32-NEXT: sw $ra, 36($sp) # 4-byte Folded Spill
+; GP32-NEXT: .cfi_offset 31, -4
+; GP32-NEXT: addu $gp, $2, $25
+; GP32-NEXT: lw $1, 60($sp)
+; GP32-NEXT: lw $2, 64($sp)
+; GP32-NEXT: lw $3, 68($sp)
+; GP32-NEXT: sw $3, 28($sp)
+; GP32-NEXT: sw $2, 24($sp)
+; GP32-NEXT: sw $1, 20($sp)
+; GP32-NEXT: lw $1, 56($sp)
+; GP32-NEXT: sw $1, 16($sp)
+; GP32-NEXT: lw $25, %call16(__divti3)($gp)
+; GP32-NEXT: jalr $25
+; GP32-NEXT: nop
+; GP32-NEXT: lw $ra, 36($sp) # 4-byte Folded Reload
+; GP32-NEXT: jr $ra
+; GP32-NEXT: addiu $sp, $sp, 40
+;
+; GP32R6-LABEL: sdiv_i128:
+; GP32R6: # %bb.0: # %entry
+; GP32R6-NEXT: lui $2, %hi(_gp_disp)
+; GP32R6-NEXT: addiu $2, $2, %lo(_gp_disp)
+; GP32R6-NEXT: addiu $sp, $sp, -40
+; GP32R6-NEXT: .cfi_def_cfa_offset 40
+; GP32R6-NEXT: sw $ra, 36($sp) # 4-byte Folded Spill
+; GP32R6-NEXT: .cfi_offset 31, -4
+; GP32R6-NEXT: addu $gp, $2, $25
+; GP32R6-NEXT: lw $1, 60($sp)
+; GP32R6-NEXT: lw $2, 64($sp)
+; GP32R6-NEXT: lw $3, 68($sp)
+; GP32R6-NEXT: sw $3, 28($sp)
+; GP32R6-NEXT: sw $2, 24($sp)
+; GP32R6-NEXT: sw $1, 20($sp)
+; GP32R6-NEXT: lw $1, 56($sp)
+; GP32R6-NEXT: sw $1, 16($sp)
+; GP32R6-NEXT: lw $25, %call16(__divti3)($gp)
+; GP32R6-NEXT: jalrc $25
+; GP32R6-NEXT: lw $ra, 36($sp) # 4-byte Folded Reload
+; GP32R6-NEXT: jr $ra
+; GP32R6-NEXT: addiu $sp, $sp, 40
+;
+; GP64-LABEL: sdiv_i128:
+; GP64: # %bb.0: # %entry
+; GP64-NEXT: daddiu $sp, $sp, -16
+; GP64-NEXT: .cfi_def_cfa_offset 16
+; GP64-NEXT: sd $ra, 8($sp) # 8-byte Folded Spill
+; GP64-NEXT: sd $gp, 0($sp) # 8-byte Folded Spill
+; GP64-NEXT: .cfi_offset 31, -8
+; GP64-NEXT: .cfi_offset 28, -16
+; GP64-NEXT: lui $1, %hi(%neg(%gp_rel(sdiv_i128)))
+; GP64-NEXT: daddu $1, $1, $25
+; GP64-NEXT: daddiu $gp, $1, %lo(%neg(%gp_rel(sdiv_i128)))
+; GP64-NEXT: ld $25, %call16(__divti3)($gp)
+; GP64-NEXT: jalr $25
+; GP64-NEXT: nop
+; GP64-NEXT: ld $gp, 0($sp) # 8-byte Folded Reload
+; GP64-NEXT: ld $ra, 8($sp) # 8-byte Folded Reload
+; GP64-NEXT: jr $ra
+; GP64-NEXT: daddiu $sp, $sp, 16
+;
+; GP64R6-LABEL: sdiv_i128:
+; GP64R6: # %bb.0: # %entry
+; GP64R6-NEXT: daddiu $sp, $sp, -16
+; GP64R6-NEXT: .cfi_def_cfa_offset 16
+; GP64R6-NEXT: sd $ra, 8($sp) # 8-byte Folded Spill
+; GP64R6-NEXT: sd $gp, 0($sp) # 8-byte Folded Spill
+; GP64R6-NEXT: .cfi_offset 31, -8
+; GP64R6-NEXT: .cfi_offset 28, -16
+; GP64R6-NEXT: lui $1, %hi(%neg(%gp_rel(sdiv_i128)))
+; GP64R6-NEXT: daddu $1, $1, $25
+; GP64R6-NEXT: daddiu $gp, $1, %lo(%neg(%gp_rel(sdiv_i128)))
+; GP64R6-NEXT: ld $25, %call16(__divti3)($gp)
+; GP64R6-NEXT: jalrc $25
+; GP64R6-NEXT: ld $gp, 0($sp) # 8-byte Folded Reload
+; GP64R6-NEXT: ld $ra, 8($sp) # 8-byte Folded Reload
+; GP64R6-NEXT: jr $ra
+; GP64R6-NEXT: daddiu $sp, $sp, 16
+;
+; MMR3-LABEL: sdiv_i128:
+; MMR3: # %bb.0: # %entry
+; MMR3-NEXT: lui $2, %hi(_gp_disp)
+; MMR3-NEXT: addiu $2, $2, %lo(_gp_disp)
+; MMR3-NEXT: addiusp -48
+; MMR3-NEXT: .cfi_def_cfa_offset 48
+; MMR3-NEXT: sw $ra, 44($sp) # 4-byte Folded Spill
+; MMR3-NEXT: swp $16, 36($sp)
+; MMR3-NEXT: .cfi_offset 31, -4
+; MMR3-NEXT: .cfi_offset 17, -8
+; MMR3-NEXT: .cfi_offset 16, -12
+; MMR3-NEXT: addu $16, $2, $25
+; MMR3-NEXT: move $1, $7
+; MMR3-NEXT: lw $7, 68($sp)
+; MMR3-NEXT: lw $17, 72($sp)
+; MMR3-NEXT: lw $3, 76($sp)
+; MMR3-NEXT: move $2, $sp
+; MMR3-NEXT: sw16 $3, 28($2)
+; MMR3-NEXT: sw16 $17, 24($2)
+; MMR3-NEXT: sw16 $7, 20($2)
+; MMR3-NEXT: lw $3, 64($sp)
+; MMR3-NEXT: sw16 $3, 16($2)
+; MMR3-NEXT: lw $25, %call16(__divti3)($16)
+; MMR3-NEXT: move $7, $1
+; MMR3-NEXT: move $gp, $16
+; MMR3-NEXT: jalr $25
+; MMR3-NEXT: nop
+; MMR3-NEXT: lwp $16, 36($sp)
+; MMR3-NEXT: lw $ra, 44($sp) # 4-byte Folded Reload
+; MMR3-NEXT: addiusp 48
+; MMR3-NEXT: jrc $ra
+;
+; MMR6-LABEL: sdiv_i128:
+; MMR6: # %bb.0: # %entry
+; MMR6-NEXT: lui $2, %hi(_gp_disp)
+; MMR6-NEXT: addiu $2, $2, %lo(_gp_disp)
+; MMR6-NEXT: addiu $sp, $sp, -48
+; MMR6-NEXT: .cfi_def_cfa_offset 48
+; MMR6-NEXT: sw $ra, 44($sp) # 4-byte Folded Spill
+; MMR6-NEXT: sw $17, 40($sp) # 4-byte Folded Spill
+; MMR6-NEXT: sw $16, 36($sp) # 4-byte Folded Spill
+; MMR6-NEXT: .cfi_offset 31, -4
+; MMR6-NEXT: .cfi_offset 17, -8
+; MMR6-NEXT: .cfi_offset 16, -12
+; MMR6-NEXT: addu $16, $2, $25
+; MMR6-NEXT: move $1, $7
+; MMR6-NEXT: lw $7, 68($sp)
+; MMR6-NEXT: lw $17, 72($sp)
+; MMR6-NEXT: lw $3, 76($sp)
+; MMR6-NEXT: move $2, $sp
+; MMR6-NEXT: sw16 $3, 28($2)
+; MMR6-NEXT: sw16 $17, 24($2)
+; MMR6-NEXT: sw16 $7, 20($2)
+; MMR6-NEXT: lw $3, 64($sp)
+; MMR6-NEXT: sw16 $3, 16($2)
+; MMR6-NEXT: lw $25, %call16(__divti3)($16)
+; MMR6-NEXT: move $7, $1
+; MMR6-NEXT: move $gp, $16
+; MMR6-NEXT: jalr $25
+; MMR6-NEXT: lw $16, 36($sp) # 4-byte Folded Reload
+; MMR6-NEXT: lw $17, 40($sp) # 4-byte Folded Reload
+; MMR6-NEXT: lw $ra, 44($sp) # 4-byte Folded Reload
+; MMR6-NEXT: addiu $sp, $sp, 48
+; MMR6-NEXT: jrc $ra
entry:
- ; ALL-LABEL: sdiv_i128:
-
- ; GP32: lw $25, %call16(__divti3)($gp)
-
- ; GP64-NOT-R6: ld $25, %call16(__divti3)($gp)
- ; 64R6: ld $25, %call16(__divti3)($gp)
-
- ; MM32: lw $25, %call16(__divti3)($16)
-
%r = sdiv i128 %a, %b
ret i128 %r
}
OpenPOWER on IntegriCloud