diff options
| author | Simon Pilgrim <llvm-dev@redking.me.uk> | 2017-08-18 14:55:50 +0000 |
|---|---|---|
| committer | Simon Pilgrim <llvm-dev@redking.me.uk> | 2017-08-18 14:55:50 +0000 |
| commit | ccaec2617577c59357c828d00068a3ee82ca2ab6 (patch) | |
| tree | 8ec945b1ca7f32dbf913463753955d82500a060c /llvm/test/CodeGen/X86 | |
| parent | 7f506f7d72d6539b4132ae38d313601a6c40e1dd (diff) | |
| download | bcm5719-llvm-ccaec2617577c59357c828d00068a3ee82ca2ab6.tar.gz bcm5719-llvm-ccaec2617577c59357c828d00068a3ee82ca2ab6.zip | |
[X86][SHA] Add scheduling latency/throughput tests for SHA instructions
llvm-svn: 311164
Diffstat (limited to 'llvm/test/CodeGen/X86')
| -rw-r--r-- | llvm/test/CodeGen/X86/sha-schedule.ll | 242 |
1 files changed, 242 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/X86/sha-schedule.ll b/llvm/test/CodeGen/X86/sha-schedule.ll new file mode 100644 index 00000000000..bf82ba8b2d9 --- /dev/null +++ b/llvm/test/CodeGen/X86/sha-schedule.ll @@ -0,0 +1,242 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=x86_64-unknown-unknown -print-schedule -mcpu=x86-64 -mattr=+sha | FileCheck %s --check-prefix=CHECK --check-prefix=GENERIC +; RUN: llc < %s -mtriple=x86_64-unknown-unknown -print-schedule -mcpu=goldmont | FileCheck %s --check-prefix=CHECK --check-prefix=GOLDMONT +; RUN: llc < %s -mtriple=x86_64-unknown-unknown -print-schedule -mcpu=cannonlake | FileCheck %s --check-prefix=CHECK --check-prefix=CANNONLAKE +; RUN: llc < %s -mtriple=x86_64-unknown-unknown -print-schedule -mcpu=znver1 | FileCheck %s --check-prefix=CHECK --check-prefix=ZNVER1 + +; +; SHA1 +; + +define <4 x i32> @test_sha1msg1(<4 x i32> %a0, <4 x i32> %a1, <4 x i32> *%a2) { +; GENERIC-LABEL: test_sha1msg1: +; GENERIC: # BB#0: +; GENERIC-NEXT: sha1msg1 %xmm1, %xmm0 +; GENERIC-NEXT: sha1msg1 (%rdi), %xmm0 +; GENERIC-NEXT: retq # sched: [1:1.00] +; +; GOLDMONT-LABEL: test_sha1msg1: +; GOLDMONT: # BB#0: +; GOLDMONT-NEXT: sha1msg1 %xmm1, %xmm0 +; GOLDMONT-NEXT: sha1msg1 (%rdi), %xmm0 +; GOLDMONT-NEXT: retq # sched: [4:1.00] +; +; CANNONLAKE-LABEL: test_sha1msg1: +; CANNONLAKE: # BB#0: +; CANNONLAKE-NEXT: sha1msg1 %xmm1, %xmm0 +; CANNONLAKE-NEXT: sha1msg1 (%rdi), %xmm0 +; CANNONLAKE-NEXT: retq # sched: [1:1.00] +; +; ZNVER1-LABEL: test_sha1msg1: +; ZNVER1: # BB#0: +; ZNVER1-NEXT: sha1msg1 %xmm1, %xmm0 +; ZNVER1-NEXT: sha1msg1 (%rdi), %xmm0 +; ZNVER1-NEXT: retq # sched: [5:0.50] + %1 = load <4 x i32>, <4 x i32>* %a2 + %2 = tail call <4 x i32> @llvm.x86.sha1msg1(<4 x i32> %a0, <4 x i32> %a1) + %3 = tail call <4 x i32> @llvm.x86.sha1msg1(<4 x i32> %2, <4 x i32> %1) + ret <4 x i32> %3 +} +declare <4 x i32> @llvm.x86.sha1msg1(<4 x i32>, <4 x i32>) + +define <4 x i32> @test_sha1msg2(<4 x i32> %a0, <4 x i32> %a1, <4 x i32> *%a2) { +; GENERIC-LABEL: test_sha1msg2: +; GENERIC: # BB#0: +; GENERIC-NEXT: sha1msg2 %xmm1, %xmm0 +; GENERIC-NEXT: sha1msg2 (%rdi), %xmm0 +; GENERIC-NEXT: retq # sched: [1:1.00] +; +; GOLDMONT-LABEL: test_sha1msg2: +; GOLDMONT: # BB#0: +; GOLDMONT-NEXT: sha1msg2 %xmm1, %xmm0 +; GOLDMONT-NEXT: sha1msg2 (%rdi), %xmm0 +; GOLDMONT-NEXT: retq # sched: [4:1.00] +; +; CANNONLAKE-LABEL: test_sha1msg2: +; CANNONLAKE: # BB#0: +; CANNONLAKE-NEXT: sha1msg2 %xmm1, %xmm0 +; CANNONLAKE-NEXT: sha1msg2 (%rdi), %xmm0 +; CANNONLAKE-NEXT: retq # sched: [1:1.00] +; +; ZNVER1-LABEL: test_sha1msg2: +; ZNVER1: # BB#0: +; ZNVER1-NEXT: sha1msg2 %xmm1, %xmm0 +; ZNVER1-NEXT: sha1msg2 (%rdi), %xmm0 +; ZNVER1-NEXT: retq # sched: [5:0.50] + %1 = load <4 x i32>, <4 x i32>* %a2 + %2 = tail call <4 x i32> @llvm.x86.sha1msg2(<4 x i32> %a0, <4 x i32> %a1) + %3 = tail call <4 x i32> @llvm.x86.sha1msg2(<4 x i32> %2, <4 x i32> %1) + ret <4 x i32> %3 +} +declare <4 x i32> @llvm.x86.sha1msg2(<4 x i32>, <4 x i32>) + +define <4 x i32> @test_sha1nexte(<4 x i32> %a0, <4 x i32> %a1, <4 x i32> *%a2) { +; GENERIC-LABEL: test_sha1nexte: +; GENERIC: # BB#0: +; GENERIC-NEXT: sha1nexte %xmm1, %xmm0 +; GENERIC-NEXT: sha1nexte (%rdi), %xmm0 +; GENERIC-NEXT: retq # sched: [1:1.00] +; +; GOLDMONT-LABEL: test_sha1nexte: +; GOLDMONT: # BB#0: +; GOLDMONT-NEXT: sha1nexte %xmm1, %xmm0 +; GOLDMONT-NEXT: sha1nexte (%rdi), %xmm0 +; GOLDMONT-NEXT: retq # sched: [4:1.00] +; +; CANNONLAKE-LABEL: test_sha1nexte: +; CANNONLAKE: # BB#0: +; CANNONLAKE-NEXT: sha1nexte %xmm1, %xmm0 +; CANNONLAKE-NEXT: sha1nexte (%rdi), %xmm0 +; CANNONLAKE-NEXT: retq # sched: [1:1.00] +; +; ZNVER1-LABEL: test_sha1nexte: +; ZNVER1: # BB#0: +; ZNVER1-NEXT: sha1nexte %xmm1, %xmm0 +; ZNVER1-NEXT: sha1nexte (%rdi), %xmm0 +; ZNVER1-NEXT: retq # sched: [5:0.50] + %1 = load <4 x i32>, <4 x i32>* %a2 + %2 = tail call <4 x i32> @llvm.x86.sha1nexte(<4 x i32> %a0, <4 x i32> %a1) + %3 = tail call <4 x i32> @llvm.x86.sha1nexte(<4 x i32> %2, <4 x i32> %1) + ret <4 x i32> %3 +} +declare <4 x i32> @llvm.x86.sha1nexte(<4 x i32>, <4 x i32>) + +define <4 x i32> @test_sha1rnds4(<4 x i32> %a0, <4 x i32> %a1, <4 x i32> *%a2) { +; GENERIC-LABEL: test_sha1rnds4: +; GENERIC: # BB#0: +; GENERIC-NEXT: sha1rnds4 $3, %xmm1, %xmm0 +; GENERIC-NEXT: sha1rnds4 $3, (%rdi), %xmm0 +; GENERIC-NEXT: retq # sched: [1:1.00] +; +; GOLDMONT-LABEL: test_sha1rnds4: +; GOLDMONT: # BB#0: +; GOLDMONT-NEXT: sha1rnds4 $3, %xmm1, %xmm0 +; GOLDMONT-NEXT: sha1rnds4 $3, (%rdi), %xmm0 +; GOLDMONT-NEXT: retq # sched: [4:1.00] +; +; CANNONLAKE-LABEL: test_sha1rnds4: +; CANNONLAKE: # BB#0: +; CANNONLAKE-NEXT: sha1rnds4 $3, %xmm1, %xmm0 +; CANNONLAKE-NEXT: sha1rnds4 $3, (%rdi), %xmm0 +; CANNONLAKE-NEXT: retq # sched: [1:1.00] +; +; ZNVER1-LABEL: test_sha1rnds4: +; ZNVER1: # BB#0: +; ZNVER1-NEXT: sha1rnds4 $3, %xmm1, %xmm0 +; ZNVER1-NEXT: sha1rnds4 $3, (%rdi), %xmm0 +; ZNVER1-NEXT: retq # sched: [5:0.50] + %1 = load <4 x i32>, <4 x i32>* %a2 + %2 = tail call <4 x i32> @llvm.x86.sha1rnds4(<4 x i32> %a0, <4 x i32> %a1, i8 3) + %3 = tail call <4 x i32> @llvm.x86.sha1rnds4(<4 x i32> %2, <4 x i32> %1, i8 3) + ret <4 x i32> %3 +} +declare <4 x i32> @llvm.x86.sha1rnds4(<4 x i32>, <4 x i32>, i8) + +; +; SHA256 +; + +define <4 x i32> @test_sha256msg1(<4 x i32> %a0, <4 x i32> %a1, <4 x i32> *%a2) { +; GENERIC-LABEL: test_sha256msg1: +; GENERIC: # BB#0: +; GENERIC-NEXT: sha256msg1 %xmm1, %xmm0 +; GENERIC-NEXT: sha256msg1 (%rdi), %xmm0 +; GENERIC-NEXT: retq # sched: [1:1.00] +; +; GOLDMONT-LABEL: test_sha256msg1: +; GOLDMONT: # BB#0: +; GOLDMONT-NEXT: sha256msg1 %xmm1, %xmm0 +; GOLDMONT-NEXT: sha256msg1 (%rdi), %xmm0 +; GOLDMONT-NEXT: retq # sched: [4:1.00] +; +; CANNONLAKE-LABEL: test_sha256msg1: +; CANNONLAKE: # BB#0: +; CANNONLAKE-NEXT: sha256msg1 %xmm1, %xmm0 +; CANNONLAKE-NEXT: sha256msg1 (%rdi), %xmm0 +; CANNONLAKE-NEXT: retq # sched: [1:1.00] +; +; ZNVER1-LABEL: test_sha256msg1: +; ZNVER1: # BB#0: +; ZNVER1-NEXT: sha256msg1 %xmm1, %xmm0 +; ZNVER1-NEXT: sha256msg1 (%rdi), %xmm0 +; ZNVER1-NEXT: retq # sched: [5:0.50] + %1 = load <4 x i32>, <4 x i32>* %a2 + %2 = tail call <4 x i32> @llvm.x86.sha256msg1(<4 x i32> %a0, <4 x i32> %a1) + %3 = tail call <4 x i32> @llvm.x86.sha256msg1(<4 x i32> %2, <4 x i32> %1) + ret <4 x i32> %3 +} +declare <4 x i32> @llvm.x86.sha256msg1(<4 x i32>, <4 x i32>) + +define <4 x i32> @test_sha256msg2(<4 x i32> %a0, <4 x i32> %a1, <4 x i32> *%a2) { +; GENERIC-LABEL: test_sha256msg2: +; GENERIC: # BB#0: +; GENERIC-NEXT: sha256msg2 %xmm1, %xmm0 +; GENERIC-NEXT: sha256msg2 (%rdi), %xmm0 +; GENERIC-NEXT: retq # sched: [1:1.00] +; +; GOLDMONT-LABEL: test_sha256msg2: +; GOLDMONT: # BB#0: +; GOLDMONT-NEXT: sha256msg2 %xmm1, %xmm0 +; GOLDMONT-NEXT: sha256msg2 (%rdi), %xmm0 +; GOLDMONT-NEXT: retq # sched: [4:1.00] +; +; CANNONLAKE-LABEL: test_sha256msg2: +; CANNONLAKE: # BB#0: +; CANNONLAKE-NEXT: sha256msg2 %xmm1, %xmm0 +; CANNONLAKE-NEXT: sha256msg2 (%rdi), %xmm0 +; CANNONLAKE-NEXT: retq # sched: [1:1.00] +; +; ZNVER1-LABEL: test_sha256msg2: +; ZNVER1: # BB#0: +; ZNVER1-NEXT: sha256msg2 %xmm1, %xmm0 +; ZNVER1-NEXT: sha256msg2 (%rdi), %xmm0 +; ZNVER1-NEXT: retq # sched: [5:0.50] + %1 = load <4 x i32>, <4 x i32>* %a2 + %2 = tail call <4 x i32> @llvm.x86.sha256msg2(<4 x i32> %a0, <4 x i32> %a1) + %3 = tail call <4 x i32> @llvm.x86.sha256msg2(<4 x i32> %2, <4 x i32> %1) + ret <4 x i32> %3 +} +declare <4 x i32> @llvm.x86.sha256msg2(<4 x i32>, <4 x i32>) + +define <4 x i32> @test_sha256rnds2(<4 x i32> %a0, <4 x i32> %a1, <4 x i32> %a2, <4 x i32> *%a3) { +; GENERIC-LABEL: test_sha256rnds2: +; GENERIC: # BB#0: +; GENERIC-NEXT: movaps %xmm0, %xmm3 # sched: [1:1.00] +; GENERIC-NEXT: movaps %xmm2, %xmm0 # sched: [1:1.00] +; GENERIC-NEXT: sha256rnds2 %xmm0, %xmm1, %xmm3 +; GENERIC-NEXT: sha256rnds2 %xmm0, (%rdi), %xmm3 +; GENERIC-NEXT: movaps %xmm3, %xmm0 # sched: [1:1.00] +; GENERIC-NEXT: retq # sched: [1:1.00] +; +; GOLDMONT-LABEL: test_sha256rnds2: +; GOLDMONT: # BB#0: +; GOLDMONT-NEXT: movaps %xmm0, %xmm3 # sched: [1:1.00] +; GOLDMONT-NEXT: movaps %xmm2, %xmm0 # sched: [1:1.00] +; GOLDMONT-NEXT: sha256rnds2 %xmm0, %xmm1, %xmm3 +; GOLDMONT-NEXT: sha256rnds2 %xmm0, (%rdi), %xmm3 +; GOLDMONT-NEXT: movaps %xmm3, %xmm0 # sched: [1:1.00] +; GOLDMONT-NEXT: retq # sched: [4:1.00] +; +; CANNONLAKE-LABEL: test_sha256rnds2: +; CANNONLAKE: # BB#0: +; CANNONLAKE-NEXT: vmovaps %xmm0, %xmm3 # sched: [1:1.00] +; CANNONLAKE-NEXT: vmovaps %xmm2, %xmm0 # sched: [1:1.00] +; CANNONLAKE-NEXT: sha256rnds2 %xmm0, %xmm1, %xmm3 +; CANNONLAKE-NEXT: sha256rnds2 %xmm0, (%rdi), %xmm3 +; CANNONLAKE-NEXT: vmovaps %xmm3, %xmm0 # sched: [1:1.00] +; CANNONLAKE-NEXT: retq # sched: [1:1.00] +; +; ZNVER1-LABEL: test_sha256rnds2: +; ZNVER1: # BB#0: +; ZNVER1-NEXT: vmovaps %xmm0, %xmm3 # sched: [1:0.50] +; ZNVER1-NEXT: vmovaps %xmm2, %xmm0 # sched: [1:0.50] +; ZNVER1-NEXT: sha256rnds2 %xmm0, %xmm1, %xmm3 +; ZNVER1-NEXT: sha256rnds2 %xmm0, (%rdi), %xmm3 +; ZNVER1-NEXT: vmovaps %xmm3, %xmm0 # sched: [1:0.50] +; ZNVER1-NEXT: retq # sched: [5:0.50] + %1 = load <4 x i32>, <4 x i32>* %a3 + %2 = tail call <4 x i32> @llvm.x86.sha256rnds2(<4 x i32> %a0, <4 x i32> %a1, <4 x i32> %a2) + %3 = tail call <4 x i32> @llvm.x86.sha256rnds2(<4 x i32> %2, <4 x i32> %1, <4 x i32> %a2) + ret <4 x i32> %3 +} +declare <4 x i32> @llvm.x86.sha256rnds2(<4 x i32>, <4 x i32>, <4 x i32>) |

