- Feb 13, 2010
-
-
Chris Lattner authored
using pred_begin/end. It is much faster. llvm-svn: 96079
-
Chris Lattner authored
instead of with pred_begin/end. llvm-svn: 96078
-
Dan Gohman authored
deterministically sorted. llvm-svn: 96071
-
Jakob Stoklund Olesen authored
Functions explicitly marked inline will get an inlining threshold slightly more aggressive than the default for -O3. This means than -O3 builds are mostly unaffected while -Os builds will be a bit bigger and faster. The difference depends entirely on how many 'inline's are sprinkled on the source. In the CINT2006 suite, only these tests are significantly affected under -Os: Size Time 471.omnetpp +1.63% -1.85% 473.astar +4.01% -6.02% 483.xalancbmk +4.60% 0.00% Note that 483.xalancbmk runs too quickly to give useful timing results. llvm-svn: 96066
-
- Feb 12, 2010
-
-
Dan Gohman authored
llvm-svn: 96005
-
Dan Gohman authored
offset distributions it doesn't expect. llvm-svn: 96002
-
Chris Lattner authored
2. don't bother trying to merge globals in non-default sections, doing so is quite dubious at best anyway. 3. fix a bug reported by Arnaud de Grandmaison where we'd try to merge two globals in different address spaces. llvm-svn: 95995
-
Daniel Dunbar authored
is breaking llvm-gcc bootstrap. llvm-svn: 95988
-
Dan Gohman authored
doesn't matter, except that ScalarEvolution tends to need less time to fold the results this way. llvm-svn: 95979
-
Dan Gohman authored
bug fixes, and with improved heuristics for analyzing foreign-loop addrecs. This change also flattens IVUsers, eliminating the stride-oriented groupings, which makes it easier to work with. llvm-svn: 95975
-
- Feb 11, 2010
-
-
Eric Christopher authored
symbols. Thanks to Duncan Sands for the testcase! llvm-svn: 95877
-
Chris Lattner authored
what it does. Enhance it to return false to optimizing vector sign extensions from vector comparisions, which is the idiom used to get a splatted vector for a vector comparison. Doing this breaks vector-casts.ll, add some compensating transformations to handle the important case they cover without depending on this canonicalization. This fixes rdar://7434900 a serious pessimization of vector compares. llvm-svn: 95855
-
Chris Lattner authored
block. Other blocks may have pointer cycles that will crash basicaa and other alias analyses. In any case, there is no point wasting cycles optimizing dead blocks. This fixes rdar://7635088 llvm-svn: 95852
-
Chris Lattner authored
instead of considering x|undef -> x, which may not be true. llvm-svn: 95850
-
Eric Christopher authored
Update testcase accordingly now that we can optimize another section. llvm-svn: 95846
-
Devang Patel authored
llvm-svn: 95828
-
- Feb 10, 2010
-
-
Devang Patel authored
llvm-svn: 95807
-
Dan Gohman authored
llvm-svn: 95781
-
- Feb 09, 2010
-
-
Eric Christopher authored
enable constant 0 offset lowering. llvm-svn: 95691
-
Eric Christopher authored
consuming for a simple optimization. llvm-svn: 95671
-
Chris Lattner authored
llvm-svn: 95643
-
Chris Lattner authored
xform. llvm-svn: 95642
-
Eric Christopher authored
llvm-svn: 95641
-
Eric Christopher authored
Initial skeleton and SCEVUnknown lowering implemented, the rest should come relatively quickly. Move testcase to new directory. Move pass to right before SimplifyLibCalls - which is moved down a bit so we can take advantage of a few opts. llvm-svn: 95628
-
Chris Lattner authored
llvm-svn: 95616
-
- Feb 06, 2010
-
-
Jakob Stoklund Olesen authored
This time it's for real! I am going to hook this up in the frontends as well. The inliner has some experimental heuristics for dealing with the inline hint. When given a -respect-inlinehint option, functions marked with the inline keyword are given a threshold just above the default for -O3. We need some experiments to determine if that is the right thing to do. llvm-svn: 95466
-
Jakob Stoklund Olesen authored
llvm-svn: 95454
-
- Feb 05, 2010
-
-
Jakob Stoklund Olesen authored
Weird code sometimes uses pointer constants other than null. This patch teaches SimplifyCFG to build switch instructions in those cases. Code like this: void f(const char *x) { if (!x) puts("null"); else if ((uintptr_t)x == 1) puts("one"); else if (x == (char*)2 || x == (char*)3) puts("two"); else if ((intptr_t)x == 4) puts("four"); else puts(x); } Now becomes a switch: define void @f(i8* %x) nounwind ssp { entry: %magicptr23 = ptrtoint i8* %x to i64 ; <i64> [#uses=1] switch i64 %magicptr23, label %if.else16 [ i64 0, label %if.then i64 1, label %if.then2 i64 2, label %if.then9 i64 3, label %if.then9 i64 4, label %if.then14 ] Note that LLVM's own DenseMap uses magic pointers. llvm-svn: 95439
-
Chris Lattner authored
xform it is checking to actually pass. There is no need to match m_SelectCst<0, -1> since instcombine canonicalizes that into not(sext). Add matches for sext(not(x)) in addition to not(sext(x)). llvm-svn: 95420
-
Dan Gohman authored
container data. This prevents it from holding onto dangling pointers and potentially behaving unpredictably. llvm-svn: 95409
-
Dan Gohman authored
malloc caller in a profile. llvm-svn: 95407
-
Eric Christopher authored
that in mind. llvm-svn: 95402
-
Bob Wilson authored
short-circuited conditions to AND/OR expressions, and those expressions are often converted back to a short-circuited form in code gen. The original source order may have been optimized to take advantage of the expected values, and if we reassociate them, we change the order and subvert that optimization. Radar 7497329. llvm-svn: 95333
-
- Feb 04, 2010
-
-
Jakob Stoklund Olesen authored
This makes the inliner about as agressive as it was before my changes to the inliner cost calculations. These levels give the same performance and slightly smaller code than before. llvm-svn: 95320
-
Eric Christopher authored
failure. llvm-svn: 95294
-
Eric Christopher authored
Fix bugs where we would compute out of bounds as in bounds, and where we couldn't know that the linker could override the size of an array. Add a few new testcases, change existing testcase to use a private global array instead of extern. llvm-svn: 95283
-
Eric Christopher authored
particular size, we just don't know what the length is yet. llvm-svn: 95266
-
- Feb 03, 2010
-
-
Bob Wilson authored
The SRThreshold value makes perfect sense for checking if an entire aggregate should be promoted to a scalar integer, but it is not so good for splitting an aggregate into its separate elements. A struct may contain a large embedded array along with some scalar fields that would benefit from being split apart by SROA. Even if the total aggregate size is large, it may still be good to perform SROA. Thus, the most important piece of this patch is simply moving the aggregate size comparison vs. SRThreshold so that it guards only the aggregate promotion. We have also been checking the number of elements to decide if an aggregate should be split up. The limit of "SRThreshold/4" seemed rather arbitrary, and I don't think it's very useful to derive this limit from SRThreshold anyway. I've collected some data showing that the current default limit of 32 (since SRThreshold defaults to 128) is a reasonable cutoff for struct types. One thing suggested by the data is that distinguishing between structs and arrays might be useful. There are (obviously) a lot more large arrays than large structs (as measured by the number of elements and not the total size -- a large array inside a struct still counts as a single element given the way we do SROA right now). Out of 8377 arrays where we successfully performed SROA while compiling a large set of benchmarks, only 16 of them had more than 8 elements. And, for those 16 arrays, it's not at all clear that SROA was actually beneficial. So, to offset the compile time cost of investigating more large structs for SROA, the patch lowers the limit on array elements to 8. This fixes Apple Radar 7563690. llvm-svn: 95224
-
Evan Cheng authored
llvm-svn: 95198
-
Bob Wilson authored
llvm-svn: 95170
-