Skip to content
  1. Sep 03, 2009
  2. Sep 02, 2009
  3. Sep 01, 2009
  4. Aug 26, 2009
  5. Aug 12, 2009
  6. Aug 11, 2009
  7. Aug 10, 2009
  8. Aug 09, 2009
  9. Aug 08, 2009
  10. Aug 06, 2009
  11. Aug 05, 2009
    • Dan Gohman's avatar
      Enable the new no-SP register classes by default. This is to address · df7ea32a
      Dan Gohman authored
      PR4572. A few tests have some minor code regressions due to different
      coalescing.
      
      llvm-svn: 78217
      df7ea32a
    • Dan Gohman's avatar
      Major calling convention code refactoring. · f9bbcd1a
      Dan Gohman authored
      Instead of awkwardly encoding calling-convention information with ISD::CALL,
      ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
      provides three virtual functions for targets to override:
      LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
      lowering done on the special nodes. They provide the same information, but
      in a more immediately usable format.
      
      This also reworks much of the target-independent tail call logic. The
      decision of whether or not to perform a tail call is now cleanly split
      between target-independent portions, and the target dependent portion
      in IsEligibleForTailCallOptimization.
      
      This also synchronizes all in-tree targets, to help enable future
      refactoring and feature work.
      
      llvm-svn: 78142
      f9bbcd1a
  12. Aug 03, 2009
  13. Aug 02, 2009
  14. Jul 30, 2009
    • Evan Cheng's avatar
      Optimize some common usage patterns of atomic built-ins __sync_add_and_fetch()... · e62288fd
      Evan Cheng authored
      Optimize some common usage patterns of atomic built-ins __sync_add_and_fetch() and __sync_sub_and_fetch. 
      
      When the return value is not used (i.e. only care about the value in the memory), x86 does not have to use add to implement these. Instead, it can use add, sub, inc, dec instructions with the "lock" prefix.
      
      This is currently implemented using a bit of instruction selection trick. The issue is the target independent pattern produces one output and a chain and we want to map it into one that just output a chain. The current trick is to select it into a merge_values with the first definition being an implicit_def. The proper solution is to add new ISD opcodes for the no-output variant. DAG combiner can then transform the node before it gets to target node selection.
      
      Problem #2 is we are adding a whole bunch of x86 atomic instructions when in fact these instructions are identical to the non-lock versions. We need a way to add target specific information to target nodes and have this information carried over to machine instructions. Asm printer (or JIT) can use this information to add the "lock" prefix.
      
      llvm-svn: 77582
      e62288fd
  15. Jul 21, 2009
    • Bill Wendling's avatar
      Add the following functions: · a8e2ad3e
      Bill Wendling authored
      - SYSCALL (normal fast system call instruction) [0f 05]
      - SYSENTER (system call entry instruction) [0f 34]
      - SYSEXIT (system call exit instruction) [0f 35]
      - SYSEXIT64 (system call exit instruction to 64-bit user code) [REX.W 0f 35]
      - SYSRET (system call return instruction) [0f 07]
      
      Patch by Sean Callanan.
      
      llvm-svn: 76528
      a8e2ad3e
  16. Jul 16, 2009
  17. Jul 14, 2009
  18. Jul 12, 2009
    • Bill Wendling's avatar
      Temporarily revert r75408. It appears to break the Apple-style builds: · 5b76fc03
      Bill Wendling authored
      x86_64-apple-darwin10-gcc -c   -g -O2  -DIN_GCC   -W -Wall -Wwrite-strings -Wstrict-prototypes -Wmissing-prototypes -pedantic -Wno-long-long -Wno-variadic-macros -Wno-overlength-strings -Wold-style-definition -Wmissing-format-attribute   -mdynamic-no-pic -DHAVE_CONFIG_H -I. -I. -I/Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmgcc42.roots/llvmgcc42~obj/src/gcc -I/Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmgcc42.roots/llvmgcc42~obj/src/gcc/. -I/Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmgcc42.roots/llvmgcc42~obj/src/gcc/../include -I./../intl -I/Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmgcc42.roots/llvmgcc42~obj/src/gcc/../libcpp/include  -I/Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmgcc42.roots/llvmgcc42~obj/src/gcc/../libdecnumber -I../libdecnumber -I/Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmCore.roots/llvmCore~dst/Developer/usr/local/include -I/Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmCore.roots/llvmCore~obj/src/include -DENABLE_LLVM -I/Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmCore.roots/llvmCore~dst/Developer/usr/local/include  -D_DEBUG  -D_GNU_SOURCE -D__STDC_LIMIT_MACROS -D__STDC_CONSTANT_MACROS -DLLVM_VERSION_INFO='"9999"' -DBUILD_LLVM_APPLE_STYLE   /Volumes/Sandbox/Buildbot/llvm/build.llvm-gcc-x86_64-darwin10-selfhost/build/llvmgcc42.roots/llvmgcc42~obj/src/gcc/tree-ssa-alias.c -o tree-ssa-alias.o
      /var/tmp//ccJQ2JBT.s:4134:Incorrect register `%rcx' used with `l' suffix
      make[2]: *** [tree-ssa-live.o] Error 1
      make[2]: *** Waiting for unfinished jobs....
      
      llvm-svn: 75412
      5b76fc03
    • Chris Lattner's avatar
      eliminate MOV64r0 in favor of a Pat<> pattern. This is only nontrivial because · 02c4339b
      Chris Lattner authored
      the div lowering code explicitly references it.
      
      llvm-svn: 75408
      02c4339b
    • Chris Lattner's avatar
      fix x86-64 static codegen to materialize the address of a global with movl instead · 38df005e
      Chris Lattner authored
      of lea.  It is better for code size (and presumably efficiency) to use:
      
        movl $foo, %eax
      
      rather than:
      
        leal foo, eax
      
      Both give a nice zero extending "move immediate" instruction, the former is just
      smaller.  Note that global addresses should be handled different by the x86
      backend, but I chose to follow the style already in place and add more fixme's.
      
      llvm-svn: 75403
      38df005e
    • Chris Lattner's avatar
      comment cleanup, reduce nesting. · 4d10f1a6
      Chris Lattner authored
      llvm-svn: 75398
      4d10f1a6
    • Chris Lattner's avatar
      remove some dead patterns, WrapperRIP doesn't exist in -static mode · 24f6ad5a
      Chris Lattner authored
      anymore, so these aren't needed.
      
      llvm-svn: 75397
      24f6ad5a
  19. Jun 27, 2009
    • Chris Lattner's avatar
      Reimplement rip-relative addressing in the X86-64 backend. The new · fea81da4
      Chris Lattner authored
      implementation primarily differs from the former in that the asmprinter
      doesn't make a zillion decisions about whether or not something will be
      RIP relative or not.  Instead, those decisions are made by isel lowering
      and propagated through to the asm printer.  To achieve this, we:
      
      1. Represent RIP relative addresses by setting the base of the X86 addr
         mode to X86::RIP.
      2. When ISel Lowering decides that it is safe to use RIP, it lowers to
         X86ISD::WrapperRIP.  When it is unsafe to use RIP, it lowers to
         X86ISD::Wrapper as before.
      3. This removes isRIPRel from X86ISelAddressMode, representing it with
         a basereg of RIP instead.
      4. The addressing mode matching logic in isel is greatly simplified.
      5. The asmprinter is greatly simplified, notably the "NotRIPRel" predicate
         passed through various printoperand routines is gone now.
      6. The various symbol printing routines in asmprinter now no longer infer
         when to emit (%rip), they just print the symbol.
      
      I think this is a big improvement over the previous situation.  It does have
      two small caveats though: 1. I implemented a horrible "no-rip" modifier for
      the inline asm "P" constraint modifier.  This is a short term hack, there is
      a much better, but more involved, solution.  2. I had to xfail an 
      -aggressive-remat testcase because it isn't handling the use of RIP in the
      constant-pool reading instruction.  This specific test is easy to fix without
      -aggressive-remat, which I intend to do next.
      
      llvm-svn: 74372
      fea81da4
  20. Jun 20, 2009
  21. Jun 16, 2009
  22. Jun 15, 2009
  23. Jun 03, 2009
  24. Jun 02, 2009
    • Evan Cheng's avatar
      448641d8
    • Dale Johannesen's avatar
      Revert 72707 and 72709, for the moment. · 5234d379
      Dale Johannesen authored
      llvm-svn: 72712
      5234d379
    • Dale Johannesen's avatar
      Make the implicit inputs and outputs of target-independent · 0b8ca792
      Dale Johannesen authored
      ADDC/ADDE use MVT::i1 (later, whatever it gets legalized to)
      instead of MVT::Flag.  Remove CARRY_FALSE in favor of 0; adjust
      all target-independent code to use this format.
      
      Most targets will still produce a Flag-setting target-dependent
      version when selection is done.  X86 is converted to use i32
      instead, which means TableGen needs to produce different code
      in xxxGenDAGISel.inc.  This keys off the new supportsHasI1 bit
      in xxxInstrInfo, currently set only for X86; in principle this
      is temporary and should go away when all other targets have
      been converted.  All relevant X86 instruction patterns are
      modified to represent setting and using EFLAGS explicitly.  The
      same can be done on other targets.
      
      The immediate behavior change is that an ADC/ADD pair are no
      longer tightly coupled in the X86 scheduler; they can be
      separated by instructions that don't clobber the flags (MOV).
      I will soon add some peephole optimizations based on using
      other instructions that set the flags to feed into ADC.
      
      llvm-svn: 72707
      0b8ca792
  25. May 31, 2009
  26. May 30, 2009
  27. May 29, 2009
  28. May 05, 2009
Loading