diff options
| author | Simon Pilgrim <llvm-dev@redking.me.uk> | 2017-10-16 21:55:09 +0000 |
|---|---|---|
| committer | Simon Pilgrim <llvm-dev@redking.me.uk> | 2017-10-16 21:55:09 +0000 |
| commit | 03c89a840a942e38127aafd7a5eb355d39994156 (patch) | |
| tree | 1f5befd49fd32116a23cb276d9a70a3fdcb8632c | |
| parent | 91259e26811fb862fa1e4dc1b463640dc24c745d (diff) | |
| download | bcm5719-llvm-03c89a840a942e38127aafd7a5eb355d39994156.tar.gz bcm5719-llvm-03c89a840a942e38127aafd7a5eb355d39994156.zip | |
[X86][3DNow] Add scheduling latency/throughput tests for 3DNow! instructions
llvm-svn: 315942
| -rw-r--r-- | llvm/test/CodeGen/X86/3dnow-schedule.ll | 372 |
1 files changed, 372 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/X86/3dnow-schedule.ll b/llvm/test/CodeGen/X86/3dnow-schedule.ll new file mode 100644 index 00000000000..c33ecc8a596 --- /dev/null +++ b/llvm/test/CodeGen/X86/3dnow-schedule.ll @@ -0,0 +1,372 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=x86_64-unknown-unknown -print-schedule -mcpu=x86-64 -mattr=+3dnowa | FileCheck %s --check-prefix=CHECK --check-prefix=GENERIC + +define void @test_femms() optsize { +; CHECK-LABEL: test_femms: +; CHECK: # BB#0: +; CHECK-NEXT: femms +; CHECK-NEXT: retq # sched: [1:1.00] + call void @llvm.x86.mmx.femms() + ret void +} +declare void @llvm.x86.mmx.femms() nounwind readnone + +define i64 @test_pavgusb(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pavgusb: +; CHECK: # BB#0: +; CHECK-NEXT: pavgusb %mm1, %mm0 +; CHECK-NEXT: pavgusb (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pavgusb(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pavgusb(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pavgusb(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pf2id(x86_mmx* %a0) optsize { +; CHECK-LABEL: test_pf2id: +; CHECK: # BB#0: +; CHECK-NEXT: pf2id (%rdi), %mm0 +; CHECK-NEXT: pf2id %mm0, %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = load x86_mmx, x86_mmx *%a0, align 8 + %2 = call x86_mmx @llvm.x86.3dnow.pf2id(x86_mmx %1) + %3 = call x86_mmx @llvm.x86.3dnow.pf2id(x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pf2id(x86_mmx) nounwind readnone + +define i64 @test_pf2iw(x86_mmx* %a0) optsize { +; CHECK-LABEL: test_pf2iw: +; CHECK: # BB#0: +; CHECK-NEXT: pf2iw (%rdi), %mm0 +; CHECK-NEXT: pf2iw %mm0, %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = load x86_mmx, x86_mmx *%a0, align 8 + %2 = call x86_mmx @llvm.x86.3dnowa.pf2iw(x86_mmx %1) + %3 = call x86_mmx @llvm.x86.3dnowa.pf2iw(x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnowa.pf2iw(x86_mmx) nounwind readnone + +define i64 @test_pfacc(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfacc: +; CHECK: # BB#0: +; CHECK-NEXT: pfacc %mm1, %mm0 +; CHECK-NEXT: pfacc (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfacc(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfacc(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfacc(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfadd(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfadd: +; CHECK: # BB#0: +; CHECK-NEXT: pfadd %mm1, %mm0 +; CHECK-NEXT: pfadd (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfadd(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfadd(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfadd(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfcmpeq(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfcmpeq: +; CHECK: # BB#0: +; CHECK-NEXT: pfcmpeq %mm1, %mm0 +; CHECK-NEXT: pfcmpeq (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfcmpeq(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfcmpeq(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfcmpeq(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfcmpge(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfcmpge: +; CHECK: # BB#0: +; CHECK-NEXT: pfcmpge %mm1, %mm0 +; CHECK-NEXT: pfcmpge (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfcmpge(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfcmpge(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfcmpge(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfcmpgt(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfcmpgt: +; CHECK: # BB#0: +; CHECK-NEXT: pfcmpgt %mm1, %mm0 +; CHECK-NEXT: pfcmpgt (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfcmpgt(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfcmpgt(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfcmpgt(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfmax(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfmax: +; CHECK: # BB#0: +; CHECK-NEXT: pfmax %mm1, %mm0 +; CHECK-NEXT: pfmax (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfmax(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfmax(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfmax(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfmin(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfmin: +; CHECK: # BB#0: +; CHECK-NEXT: pfmin %mm1, %mm0 +; CHECK-NEXT: pfmin (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfmin(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfmin(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfmin(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfmul(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfmul: +; CHECK: # BB#0: +; CHECK-NEXT: pfmul %mm1, %mm0 +; CHECK-NEXT: pfmul (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfmul(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfmul(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfmul(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfnacc(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfnacc: +; CHECK: # BB#0: +; CHECK-NEXT: pfnacc %mm1, %mm0 +; CHECK-NEXT: pfnacc (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnowa.pfnacc(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnowa.pfnacc(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnowa.pfnacc(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfpnacc(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfpnacc: +; CHECK: # BB#0: +; CHECK-NEXT: pfpnacc %mm1, %mm0 +; CHECK-NEXT: pfpnacc (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnowa.pfpnacc(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnowa.pfpnacc(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnowa.pfpnacc(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfrcp(x86_mmx* %a0) optsize { +; CHECK-LABEL: test_pfrcp: +; CHECK: # BB#0: +; CHECK-NEXT: pfrcp (%rdi), %mm0 +; CHECK-NEXT: pfrcp %mm0, %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = load x86_mmx, x86_mmx *%a0, align 8 + %2 = call x86_mmx @llvm.x86.3dnow.pfrcp(x86_mmx %1) + %3 = call x86_mmx @llvm.x86.3dnow.pfrcp(x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfrcp(x86_mmx) nounwind readnone + +define i64 @test_pfrcpit1(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfrcpit1: +; CHECK: # BB#0: +; CHECK-NEXT: pfrcpit1 %mm1, %mm0 +; CHECK-NEXT: pfrcpit1 (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfrcpit1(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfrcpit1(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfrcpit1(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfrcpit2(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfrcpit2: +; CHECK: # BB#0: +; CHECK-NEXT: pfrcpit2 %mm1, %mm0 +; CHECK-NEXT: pfrcpit2 (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfrcpit2(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfrcpit2(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfrcpit2(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfrsqit1(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfrsqit1: +; CHECK: # BB#0: +; CHECK-NEXT: pfrsqit1 %mm1, %mm0 +; CHECK-NEXT: pfrsqit1 (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfrsqit1(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfrsqit1(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfrsqit1(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfrsqrt(x86_mmx* %a0) optsize { +; CHECK-LABEL: test_pfrsqrt: +; CHECK: # BB#0: +; CHECK-NEXT: pfrsqrt (%rdi), %mm0 +; CHECK-NEXT: pfrsqrt %mm0, %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = load x86_mmx, x86_mmx *%a0, align 8 + %2 = call x86_mmx @llvm.x86.3dnow.pfrsqrt(x86_mmx %1) + %3 = call x86_mmx @llvm.x86.3dnow.pfrsqrt(x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfrsqrt(x86_mmx) nounwind readnone + +define i64 @test_pfsub(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfsub: +; CHECK: # BB#0: +; CHECK-NEXT: pfsub %mm1, %mm0 +; CHECK-NEXT: pfsub (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfsub(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfsub(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfsub(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pfsubr(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pfsubr: +; CHECK: # BB#0: +; CHECK-NEXT: pfsubr %mm1, %mm0 +; CHECK-NEXT: pfsubr (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pfsubr(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pfsubr(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pfsubr(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pi2fd(x86_mmx* %a0) optsize { +; CHECK-LABEL: test_pi2fd: +; CHECK: # BB#0: +; CHECK-NEXT: pi2fd (%rdi), %mm0 +; CHECK-NEXT: pi2fd %mm0, %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = load x86_mmx, x86_mmx *%a0, align 8 + %2 = call x86_mmx @llvm.x86.3dnow.pi2fd(x86_mmx %1) + %3 = call x86_mmx @llvm.x86.3dnow.pi2fd(x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pi2fd(x86_mmx) nounwind readnone + +define i64 @test_pi2fw(x86_mmx* %a0) optsize { +; CHECK-LABEL: test_pi2fw: +; CHECK: # BB#0: +; CHECK-NEXT: pi2fw (%rdi), %mm0 +; CHECK-NEXT: pi2fw %mm0, %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = load x86_mmx, x86_mmx *%a0, align 8 + %2 = call x86_mmx @llvm.x86.3dnowa.pi2fw(x86_mmx %1) + %3 = call x86_mmx @llvm.x86.3dnowa.pi2fw(x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnowa.pi2fw(x86_mmx) nounwind readnone + +define i64 @test_pmulhrw(x86_mmx %a0, x86_mmx %a1, x86_mmx* %a2) optsize { +; CHECK-LABEL: test_pmulhrw: +; CHECK: # BB#0: +; CHECK-NEXT: pmulhrw %mm1, %mm0 +; CHECK-NEXT: pmulhrw (%rdi), %mm0 +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = call x86_mmx @llvm.x86.3dnow.pmulhrw(x86_mmx %a0, x86_mmx %a1) + %2 = load x86_mmx, x86_mmx *%a2, align 8 + %3 = call x86_mmx @llvm.x86.3dnow.pmulhrw(x86_mmx %1, x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnow.pmulhrw(x86_mmx, x86_mmx) nounwind readnone + +define i64 @test_pswapd(x86_mmx* %a0) optsize { +; CHECK-LABEL: test_pswapd: +; CHECK: # BB#0: +; CHECK-NEXT: pswapd (%rdi), %mm0 # mm0 = mem[1,0] +; CHECK-NEXT: pswapd %mm0, %mm0 # mm0 = mm0[1,0] +; CHECK-NEXT: movd %mm0, %rax # sched: [1:0.33] +; CHECK-NEXT: retq # sched: [1:1.00] + %1 = load x86_mmx, x86_mmx *%a0, align 8 + %2 = call x86_mmx @llvm.x86.3dnowa.pswapd(x86_mmx %1) + %3 = call x86_mmx @llvm.x86.3dnowa.pswapd(x86_mmx %2) + %4 = bitcast x86_mmx %3 to i64 + ret i64 %4 +} +declare x86_mmx @llvm.x86.3dnowa.pswapd(x86_mmx) nounwind readnone |

