summaryrefslogtreecommitdiffstats
path: root/llvm/test/CodeGen/X86/tlv-1.ll
diff options
context:
space:
mode:
authorEvan Cheng <evan.cheng@apple.com>2011-01-07 19:35:30 +0000
committerEvan Cheng <evan.cheng@apple.com>2011-01-07 19:35:30 +0000
commita048c83fe476e47bd549d70f423e1949a520f7af (patch)
treef77e6756fe59890824ef9aec93e0b43cf256aa0b /llvm/test/CodeGen/X86/tlv-1.ll
parent2cd32a025139abd759ca51e970c8ad10310e2137 (diff)
downloadbcm5719-llvm-a048c83fe476e47bd549d70f423e1949a520f7af.tar.gz
bcm5719-llvm-a048c83fe476e47bd549d70f423e1949a520f7af.zip
Revert r122955. It seems using movups to lower memcpy can cause massive regression (even on Nehalem) in edge cases. I also didn't see any real performance benefit.
llvm-svn: 123015
Diffstat (limited to 'llvm/test/CodeGen/X86/tlv-1.ll')
-rw-r--r--llvm/test/CodeGen/X86/tlv-1.ll8
1 files changed, 2 insertions, 6 deletions
diff --git a/llvm/test/CodeGen/X86/tlv-1.ll b/llvm/test/CodeGen/X86/tlv-1.ll
index f9be15d1a95..42940f147ed 100644
--- a/llvm/test/CodeGen/X86/tlv-1.ll
+++ b/llvm/test/CodeGen/X86/tlv-1.ll
@@ -10,12 +10,8 @@ entry:
unreachable
; CHECK: movq _c@TLVP(%rip), %rdi
; CHECK-NEXT: callq *(%rdi)
- ; CHECK-NEXT: pxor %xmm0, %xmm0
- ; CHECK-NEXT: movups %xmm0, 32(%rax)
- ; CHECK-NEXT: movups %xmm0, 16(%rax)
- ; CHECK-NEXT: movups %xmm0, (%rax)
- ; CHECK-NEXT: movl $0, 56(%rax)
- ; CHECK-NEXT: movq $0, 48(%rax)
+ ; CHECK-NEXT: movl $0, 56(%rax)
+ ; CHECK-NEXT: movq $0, 48(%rax)
}
declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i32, i1) nounwind
OpenPOWER on IntegriCloud