- Mar 04, 2012
-
-
Chandler Carruth authored
llvm-svn: 152000
-
Chandler Carruth authored
llvm-svn: 151999
-
Craig Topper authored
llvm-svn: 151998
-
Craig Topper authored
llvm-svn: 151996
-
Craig Topper authored
Use uint8_t instead of enums to store values in X86 disassembler table. Shaves 150k off the size of X86DisassemblerDecoder.o llvm-svn: 151995
-
- Mar 03, 2012
-
-
Benjamin Kramer authored
llvm-svn: 151986
-
Benjamin Kramer authored
- Shrink the opcode field to 16 bits. - Shrink the AsmVariantID field to 8 bits. - Store the mnemonic string in a string table, store a 16 bit index. - Store a pascal-style length byte in the string instead of a null terminator, so we can avoid calling strlen on every entry we visit during mnemonic search. Shrinks X86AsmParser.o from 434k to 201k on x86_64 and eliminates relocs from the table. llvm-svn: 151984
-
Benjamin Kramer authored
StringToOffsetTable: Allow uniquing the first element, add an option to skip appending a terminating null. llvm-svn: 151983
-
Benjamin Kramer authored
While at it bump the small vector size a bit, it's inside a heap-allocated class. llvm-svn: 151980
-
Rafael Espindola authored
llvm-svn: 151979
-
Duncan Sands authored
llvm-svn: 151977
-
Francois Pichet authored
Move the NonPOD struct out of the anonymous namespace instead of adding llvm:: everywhere to fix the HashingTest on MSVC . chandlerc proposed this better solution on IRC. llvm-svn: 151974
-
Duncan Sands authored
llvm-svn: 151973
-
Francois Pichet authored
llvm-svn: 151971
-
NAKAMURA Takumi authored
unittests/ADT/HashingTest.cpp: Temporarily disable a new test introduced in r151891, to appease msvc. llvm-svn: 151970
-
NAKAMURA Takumi authored
llvm-svn: 151968
-
Daniel Dunbar authored
llvm-svn: 151959
-
Jakob Stoklund Olesen authored
llvm-svn: 151958
-
David Meyer authored
llvm-svn: 151957
-
- Mar 02, 2012
-
-
Benjamin Kramer authored
llvm-svn: 151932
-
Jim Grosbach authored
llvm-svn: 151926
-
Daniel Dunbar authored
llvm-svn: 151921
-
Benjamin Kramer authored
This could probably be made a lot smarter, but this is a common case and doesn't require LVI to scan a lot of code. With this change CVP can optimize away the "shift == 0" case in Hashing.h that only gets hit when "shift" is in a range not containing 0. llvm-svn: 151919
-
Benjamin Kramer authored
Hashing: microoptimize a truncate on 64 bit away. This currently blocks dead code eliminating the conditional. The optimizer should handle this eventually, but currently LVI isn't really designed for this kind of stuff. llvm-svn: 151918
-
Jia Liu authored
llvm-svn: 151909
-
Jia Liu authored
llvm-svn: 151908
-
Chandler Carruth authored
folks who know something about PPC tell me that the byte swap is crazy fast and without this the bit mixture would actually be different. It might not be worse, but I've not measured it and so I'd rather not trust it. This way, the algorithm is identical on both endianness hosts. I'll look into any performance issues etc stemming from this. llvm-svn: 151892
-
Chandler Carruth authored
just ensure that the number of bytes in the pair is the sum of the bytes in each side of the pair. As long as thats true, there are no extra bytes that might be padding. Also add a few tests that previously would have slipped through the checking. The more accurate checking mechanism catches these and ensures they are handled conservatively correctly. Thanks to Duncan for prodding me to do this right and more simply. llvm-svn: 151891
-
Evgeniy Stepanov authored
This change replaces getTypeStoreSize with getTypeAllocSize in AddressSanitizer instrumentation for stack allocations. One case where old behaviour produced undesired results is an optimization in InstCombine pass (PromoteCastOfAllocation), which can replace alloca(T) with alloca(S), where S has the same AllocSize, but a smaller StoreSize. Another case is memcpy(long double => long double), where ASan will poison bytes 10-15 of a stack-allocated long double (StoreSize 10, AllocSize 16, sizeof(long double) = 16). See http://llvm.org/bugs/show_bug.cgi?id=12047 for more context. llvm-svn: 151887
-
Chandler Carruth authored
llvm-svn: 151886
-
Chandler Carruth authored
for 32-bit builds in here. llvm-svn: 151885
-
Chandler Carruth authored
offsetof buildbot errors to go away... llvm-svn: 151884
-
Chandler Carruth authored
hashable data. This matters when we have pair<T*, U*> as a key, which is quite common in DenseMap, etc. To that end, we need to detect when this is safe. The requirements on a generic std::pair<T, U> are: 1) Both T and U must satisfy the existing is_hashable_data trait. Note that this includes the requirement that T and U have no internal padding bits or other bits not contributing directly to equality. 2) The alignment constraints of std::pair<T, U> do not require padding between consecutive objects. 3) The alignment constraints of U and the size of T do not conspire to require padding between the first and second elements. Grow two somewhat magical traits to detect this by forming a pod structure and inspecting offset artifacts on it. Hopefully this won't cause any compilers to panic. Added and adjusted tests now that pairs, even nested pairs, are treated as just sequences of data. Thanks to Jeffrey Yasskin for helping me sort through this and reviewing the somewhat subtle traits. llvm-svn: 151883
-
Chandler Carruth authored
an open question of whether we can do better than this by treating pairs as boring data containers and directly hashing the two subobjects. This at least makes the API reasonable. In order to make this change, I reorganized the header a bit. I lifted the declarations of the hash_value functions up to the top of the header with their doxygen comments as these are intended for users to interact with. They shouldn't have to wade through implementation details. I then defined them at the very end so that they could be defined in terms of hash_combine or any other hashing infrastructure. Added various pair-hashing unittests. llvm-svn: 151882
-
Chad Rosier authored
In this instance we are generating the tail-call during legalizeDAG. The 2nd floor call can't be a tail call because it clobbers %xmm1, which is defined by the first floor call. The first floor call can't be a tail-call because it's not in the tail position. The only reasonable way I could think to fix this in a target-independent manner was to check for glue logic on the copy reg. rdar://10930395 llvm-svn: 151877
-
Eric Christopher authored
llvm-svn: 151875
-
Eric Christopher authored
llvm-svn: 151874
-
Eric Christopher authored
to the string table for the function name, not the function name. llvm-svn: 151873
-
Dan Gohman authored
can insert a new element, invalidating iterators. Use find instead, and handle the case where the key is not found explicitly. llvm-svn: 151871
-
Dan Gohman authored
llvm-svn: 151869
-