- Oct 29, 2007
-
-
Dan Gohman authored
llvm-svn: 43470
-
Dan Gohman authored
lowering load and store instructions. llvm-svn: 43468
-
Dan Gohman authored
llvm-svn: 43467
-
Dan Gohman authored
of just printing to cerr. llvm-svn: 43466
-
Evan Cheng authored
transformation. Previously, it's restricted by ensuring the number of load uses is one. Now the restriction is loosened up by allowing setcc uses to be "extended" (e.g. setcc x, c, eq -> setcc sext(x), sext(c), eq). llvm-svn: 43465
-
Dan Gohman authored
llvm-svn: 43464
-
Dan Gohman authored
llvm-svn: 43463
-
Dan Gohman authored
llvm-svn: 43462
-
Dan Gohman authored
llvm-svn: 43461
-
Dan Gohman authored
llvm-svn: 43460
-
Ted Kremenek authored
constant to an unsigned int. We now just directly assign the literal 0. llvm-svn: 43459
-
Evan Cheng authored
llvm-svn: 43446
-
Chris Lattner authored
llvm-svn: 43444
-
Chris Lattner authored
now. It conflicts with clang's -pedantic flag. llvm-svn: 43431
-
Chris Lattner authored
b/h/w/k/q inline asm memory modifiers, which are just ignored. This fixes PR1748 and CodeGen/X86/2007-10-28-inlineasm-q-modifier.ll llvm-svn: 43430
-
Chris Lattner authored
zero-length fields better. llvm-svn: 43427
-
Chris Lattner authored
can have uses too. Wouldn't it be nice if invoke didn't exist? :) llvm-svn: 43426
-
Ted Kremenek authored
pointers that were not backpatched (previously checked the wrong invariant). llvm-svn: 43425
-
- Oct 28, 2007
-
-
Anton Korobeynikov authored
llvm-svn: 43424
-
Ted Kremenek authored
eager backpatching instead of waithing until all objects have been deserialized. This allows us to reduce the memory footprint needed for backpatching. llvm-svn: 43422
-
Duncan Sands authored
of offset and the alignment of ptr if these are both powers of 2. While the ptr alignment is guaranteed to be a power of 2, there is no reason to think that offset is. For example, if offset is 12 (the size of a long double on x86-32 linux) and the alignment of ptr is 8, then the alignment of ptr+offset will in general be 4, not 8. Introduce a function MinAlign, lifted from gcc, for computing the minimum guaranteed alignment. I've tried to fix up everywhere under lib/CodeGen/SelectionDAG/. I also changed some places that weren't wrong (because both values were a power of 2), as a defensive change against people copying and pasting the code. Hopefully someone who cares about alignment will review the rest of LLVM and fix up the remaining places. Since I'm on x86 I'm not very motivated to do this myself... llvm-svn: 43421
-
Evan Cheng authored
llvm-svn: 43420
-
- Oct 27, 2007
-
-
Evan Cheng authored
- ChangeCompareStride only reuse stride that is larger than current stride. It will let the general reuse mechanism to try to reuse a smaller stride. - Watch out for multiplication overflow in ChangeCompareStride. - Replace std::set with SmallPtrSet. llvm-svn: 43408
-
- Oct 26, 2007
-
-
Ted Kremenek authored
llvm-svn: 43405
-
Bill Wendling authored
FE. - Explicitly pass in the alignment of the load & store. - XFAIL 2007-10-23-UnalignedMemcpy.ll because llc has a bug that crashes on unaligned pointers. llvm-svn: 43398
-
Evan Cheng authored
llvm-svn: 43384
-
Anton Korobeynikov authored
registers in case, when FP pointer was eliminated. This should fixes misc. random EH-related crahses, when stuff is compiled with -fomit-frame-pointer. Thanks Duncan for nailing this bug! llvm-svn: 43381
-
Eric Christopher authored
want them later (mips32/64). llvm-svn: 43380
-
Owen Anderson authored
llvm-svn: 43379
-
Gordon Henriksen authored
improved wording in source files. llvm-svn: 43377
-
Evan Cheng authored
Loosen up iv reuse to allow reuse of the same stride but a larger type when truncating from the larger type to smaller type is free. e.g. Turns this loop: LBB1_1: # entry.bb_crit_edge xorl %ecx, %ecx xorw %dx, %dx movw %dx, %si LBB1_2: # bb movl L_X$non_lazy_ptr, %edi movw %si, (%edi) movl L_Y$non_lazy_ptr, %edi movw %dx, (%edi) addw $4, %dx incw %si incl %ecx cmpl %eax, %ecx jne LBB1_2 # bb into LBB1_1: # entry.bb_crit_edge xorl %ecx, %ecx xorw %dx, %dx LBB1_2: # bb movl L_X$non_lazy_ptr, %esi movw %cx, (%esi) movl L_Y$non_lazy_ptr, %esi movw %dx, (%esi) addw $4, %dx incl %ecx cmpl %eax, %ecx jne LBB1_2 # bb llvm-svn: 43375
-
Hartmut Kaiser authored
llvm-svn: 43374
-
Ted Kremenek authored
pointers that employ unused bits in a pointer to store extra data. llvm-svn: 43373
-
Hartmut Kaiser authored
Silenced VC++ warning. llvm-svn: 43372
-
Hartmut Kaiser authored
llvm-svn: 43369
-
Evan Cheng authored
stride may be rewritten using the stride of the compare instruction. llvm-svn: 43367
-
- Oct 25, 2007
-
-
Dale Johannesen authored
llvm-svn: 43364
-
Bill Wendling authored
llvm-svn: 43359
-
Ted Kremenek authored
llvm-svn: 43357
-
Evan Cheng authored
llvm-svn: 43356
-