- Dec 06, 2011
-
-
Lang Hames authored
llvm-svn: 145893
-
Chad Rosier authored
rdar://10528060 llvm-svn: 145891
-
Jakob Stoklund Olesen authored
Previously, all ARM::CONSTPOOL_ENTRY instructions had a hardwired alignment of 4 bytes emitted by ARMAsmPrinter. Now the same alignment is set on the basic block. This is in preparation of supporting ARM constant pool islands with different alignments. llvm-svn: 145890
-
Jakob Stoklund Olesen authored
This was actually a bit of a mess. TLI.setPrefLoopAlignment was clearly documented as taking log2(bytes) units, but the x86 target would still set a preferred loop alignment of '16'. CodePlacementOpt passed this number on to the basic block, and AsmPrinter interpreted it as bytes. Now both MachineFunction and MachineBasicBlock use logarithmic alignments. Obviously, MachineConstantPool still measures alignments in bytes, so we can emulate the thrill of using as. llvm-svn: 145889
-
Bill Wendling authored
value over that much. llvm-svn: 145888
-
Jim Grosbach authored
rdar://10069056 llvm-svn: 145885
-
Jakob Stoklund Olesen authored
llvm-svn: 145883
-
Jim Grosbach authored
Whether a fixup needs relaxation for the associated instruction is a target-specific function, as the FIXME indicated. Create a hook for that and use it. llvm-svn: 145881
-
Nick Lewycky authored
llvm-svn: 145880
-
Chad Rosier authored
don't do this now, but add a test case to prevent this from happening in the future. Additional test for rdar://9892684 llvm-svn: 145879
-
Jim Grosbach authored
llvm-svn: 145878
-
Jim Grosbach authored
per http://llvm.org/docs/CodingStandards.html#ll_naming llvm-svn: 145873
-
Jim Grosbach authored
Not right yet, as the rules for when to relax in the MCAssembler aren't (yet) correct for ARM. This is a step in the proper direction, though. llvm-svn: 145871
-
Nick Lewycky authored
memory fences) in statistics registration, which works the same way that ManagedStatic registration does. llvm-svn: 145869
-
- Dec 05, 2011
-
-
Chad Rosier authored
llvm-svn: 145866
-
Chad Rosier authored
where this would be bad as the backend shouldn't have a problem inlining small memcpys. rdar://10510150 llvm-svn: 145865
-
Jim Grosbach authored
llvm-svn: 145863
-
Jim Grosbach authored
rdar://10529664 llvm-svn: 145860
-
Akira Hatanaka authored
PerformANDCombine and PerformOrCombine aware of them. Test cases are included too. llvm-svn: 145853
-
Akira Hatanaka authored
them. llvm-svn: 145852
-
Jim Grosbach authored
rdar://10529348 llvm-svn: 145851
-
Akira Hatanaka authored
O32 with relocation-model=pic too. llvm-svn: 145850
-
Jim Grosbach authored
Finish up rdar://10522016. llvm-svn: 145846
-
Jim Grosbach authored
llvm-svn: 145844
-
Jim Grosbach authored
llvm-svn: 145843
-
Jim Grosbach authored
Combined destination and first source operand for f32 variant of the VMUL (by scalar) instruction. rdar://10522016 llvm-svn: 145842
-
Hal Finkel authored
llvm-svn: 145819
-
Hal Finkel authored
llvm-svn: 145818
-
Hal Finkel authored
llvm-svn: 145816
-
Benjamin Kramer authored
- Calling getUser in a loop is much more expensive than iterating over a few instructions. - Use it instead of the open-coded loop in AddrModeMatcher. - 5% speedup on ARMDisassembler.cpp Release builds. llvm-svn: 145810
-
Craig Topper authored
llvm-svn: 145804
-
Craig Topper authored
llvm-svn: 145803
-
Nadav Rotem authored
Add support for vectors of pointers. llvm-svn: 145801
-
- Dec 04, 2011
-
-
Eric Christopher authored
not get there any other way. llvm-svn: 145789
-
Bob Wilson authored
llvm-svn: 145783
-
Anton Korobeynikov authored
Maybe some targets should use this as well. Patch by Evgeniy Stepanov! llvm-svn: 145781
-
- Dec 03, 2011
-
-
Venkatraman Govindaraju authored
AnalyzeBranch doesn't change the successor, just the order. llvm-svn: 145779
-
Benjamin Kramer authored
-3% on ARMDissasembler.cpp. llvm-svn: 145773
-
Benjamin Kramer authored
llvm-svn: 145771
-
Benjamin Kramer authored
Add a "seen blocks" cache to LVI to avoid a linear scan over the whole cache just to remove no blocks from the maps. -15% on ARMDisassembler.cpp (Release build). It's not that great to add another layer of caching to the caching-heavy LVI but I don't see a better way. llvm-svn: 145770
-