| Commit message (Collapse) | Author | Age | Files | Lines |
| ... | |
| |
|
|
| |
llvm-svn: 147725
|
| |
|
|
|
|
| |
This will be more important as we extend the LSR pass in ways that don't rely on the formula solver. In particular, we need it for constructing IV chains.
llvm-svn: 147724
|
| |
|
|
|
|
|
|
|
| |
This enables basic local CSE, giving us 20% smaller code for
consumer-typeset in -O0 builds.
<rdar://problem/10658692>
llvm-svn: 147720
|
| |
|
|
|
|
|
|
|
|
| |
up so branch folding pass can't use the scavenger. :-( This doesn't breaks
anything currently. It just means targets which do not carefully update kill
markers cannot run post-ra scheduler (not new, it has always been the case).
We should fix this at some point since it's really hacky.
llvm-svn: 147719
|
| |
|
|
|
|
|
|
| |
LoopSimplify may not run on some outer loops, e.g. because of indirect
branches. SCEVExpander simply cannot handle outer loops with no preheaders.
Fixes rdar://10655343 SCEVExpander segfault.
llvm-svn: 147718
|
| |
|
|
|
|
|
|
| |
file error checking. Use that to error on an unfinished cfi_startproc.
The error is not nice, but is already better than a segmentation fault.
llvm-svn: 147717
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
opportunities that only present themselves after late optimizations
such as tail duplication .e.g.
## BB#1:
movl %eax, %ecx
movl %ecx, %eax
ret
The register allocator also leaves some of them around (due to false
dep between copies from phi-elimination, etc.)
This required some changes in codegen passes. Post-ra scheduler and the
pseudo-instruction expansion passes have been moved after branch folding
and tail merging. They were before branch folding before because it did
not always update block livein's. That's fixed now. The pass change makes
independently since we want to properly schedule instructions after
branch folding / tail duplication.
rdar://10428165
rdar://10640363
llvm-svn: 147716
|
| |
|
|
|
|
|
| |
exposed with an upcoming change will would delete the copy to return register
because there is no use! It's amazing anything works.
llvm-svn: 147715
|
| |
|
|
|
|
|
|
|
|
|
|
|
| |
This eliminates a lot of constant pool entries for -O0 builds of code
with many global variable accesses.
This speeds up -O0 codegen of consumer-typeset by 2x because the
constant island pass no longer has to look at thousands of constant pool
entries.
<rdar://problem/10629774>
llvm-svn: 147712
|
| |
|
|
| |
llvm-svn: 147711
|
| |
|
|
|
|
| |
replaceCongruentPhis checkin.
llvm-svn: 147709
|
| |
|
|
| |
llvm-svn: 147707
|
| |
|
|
|
|
| |
Fixes rdar://10614894
llvm-svn: 147704
|
| |
|
|
| |
llvm-svn: 147703
|
| |
|
|
| |
llvm-svn: 147700
|
| |
|
|
| |
llvm-svn: 147696
|
| |
|
|
|
|
| |
to bleed from the eyes.
llvm-svn: 147695
|
| |
|
|
| |
llvm-svn: 147694
|
| |
|
|
| |
llvm-svn: 147693
|
| |
|
|
|
|
| |
Experiments show this to be a small speedup for modern ARM cores.
llvm-svn: 147689
|
| |
|
|
|
|
| |
isIVUserOrOperand.
llvm-svn: 147686
|
| |
|
|
| |
llvm-svn: 147685
|
| |
|
|
| |
llvm-svn: 147683
|
| |
|
|
| |
llvm-svn: 147682
|
| |
|
|
| |
llvm-svn: 147679
|
| |
|
|
| |
llvm-svn: 147676
|
| |
|
|
| |
llvm-svn: 147675
|
| |
|
|
|
|
|
|
| |
lldb testsuite.
rdar://10652330
llvm-svn: 147673
|
| |
|
|
| |
llvm-svn: 147667
|
| |
|
|
| |
llvm-svn: 147654
|
| |
|
|
|
|
|
|
|
| |
the debug type accelerator tables to contain the tag and a flag
stating whether or not a compound type is a complete type.
rdar://10652330
llvm-svn: 147651
|
| |
|
|
|
|
|
|
|
|
|
|
| |
present in the bottom of the CFG triangle, as the transformation isn't
ever valuable if the branch can't be eliminated.
Also, unify some heuristics between SimplifyCFG's multiple
if-converters, for consistency.
This fixes rdar://10627242.
llvm-svn: 147630
|
| |
|
|
|
|
| |
global initializers if there's an implied extension or truncation.
llvm-svn: 147625
|
| |
|
|
|
|
|
|
| |
System V Application Binary Interface. This lets us use
-fvisibility-inlines-hidden with LTO.
Fixes PR11697.
llvm-svn: 147624
|
| |
|
|
|
|
|
|
| |
code can incorrectly move the load across a store. This never
happens in practice today, but only because the current
heuristics accidentally preclude it.
llvm-svn: 147623
|
| |
|
|
| |
llvm-svn: 147618
|
| |
|
|
|
|
| |
Eliminate the dead test for it on each loop iteration. No functionality change.
llvm-svn: 147616
|
| |
|
|
| |
llvm-svn: 147615
|
| |
|
|
|
|
| |
called with the base class.
llvm-svn: 147610
|
| |
|
|
| |
llvm-svn: 147608
|
| |
|
|
| |
llvm-svn: 147605
|
| |
|
|
|
|
|
|
|
| |
a combined-away node and the result of the combine isn't substantially
smaller than the input, it's just canonicalized. This is the first part
of a significant (7%) performance gain for Snappy's hot decompression
loop.
llvm-svn: 147604
|
| |
|
|
| |
llvm-svn: 147602
|
| |
|
|
|
|
|
|
|
| |
instruction combining of sequences generated by ptestz/ptestc intrinsics to ptest+jcc pair for SSE and AVX.
Testing: passed 'make check' including LIT tests for all sequences being handled (both SSE and AVX)
Reviewers: Evan Cheng, David Blaikie, Bruno Lopes, Elena Demikhovsky, Chad Rosier, Anton Korobeynikov
llvm-svn: 147601
|
| |
|
|
|
|
| |
antidependence latency on ARM in exceedingly rare cases.
llvm-svn: 147594
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
efficient.
This small bit of ASM code is sufficient to do what the old algorithm did:
movq %rax, %xmm0
punpckldq (c0), %xmm0 // c0: (uint4){ 0x43300000U, 0x45300000U, 0U, 0U }
subpd (c1), %xmm0 // c1: (double2){ 0x1.0p52, 0x1.0p52 * 0x1.0p32 }
#ifdef __SSE3__
haddpd %xmm0, %xmm0
#else
pshufd $0x4e, %xmm0, %xmm1
addpd %xmm1, %xmm0
#endif
It's arguably faster. One caveat, the 'haddpd' instruction isn't very fast on
all processors.
<rdar://problem/7719814>
llvm-svn: 147593
|
| |
|
|
|
|
|
|
|
|
|
|
| |
Now that canRealignStack() understands frozen reserved registers, it is
safe to use it for aligned spill instructions.
It will only return true if the registers reserved at the beginning of
register allocation allow for dynamic stack realignment.
<rdar://problem/10625436>
llvm-svn: 147579
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Once register allocation has started the reserved registers are frozen.
Fix the ARM canRealignStack() hook to respect the frozen register state.
Now the hook returns false if register allocation was started with frame
pointer elimination enabled.
It also returns false if register allocation started without a reserved
base pointer, and stack realignment would require a base pointer. This
bug was breaking oggenc on armv6.
No test case, an upcoming patch will use this functionality to realign
the stack for spill slots when possible.
llvm-svn: 147578
|
| |
|
|
|
|
|
|
|
|
|
| |
The register allocators don't currently support adding reserved
registers while they are running. Extend the MRI API to keep track of
the set of reserved registers when register allocation started.
Target hooks like hasFP() and needsStackRealignment() can look at this
set to avoid reserving more registers during register allocation.
llvm-svn: 147577
|
| |
|
|
|
|
|
| |
Values, rather than just Instructions, since it's interesting
for ConstantExprs too.
llvm-svn: 147560
|