Skip to content
  1. Nov 05, 2008
    • Evan Cheng's avatar
      Add more vector move low and zero-extend patterns. · 27889ab2
      Evan Cheng authored
      llvm-svn: 58752
      27889ab2
    • Evan Cheng's avatar
      Indentation. · 3cd5e8c9
      Evan Cheng authored
      llvm-svn: 58750
      3cd5e8c9
    • Dan Gohman's avatar
      Eliminate the ISel priority queue, which used the topological order for a · f14b77eb
      Dan Gohman authored
      priority function. Instead, just iterate over the AllNodes list, which is
      already in topological order. This eliminates a fair amount of bookkeeping,
      and speeds up the isel phase by about 15% on many testcases.
      
      The impact on most targets is that AddToISelQueue calls can be simply removed.
      
      In the x86 target, there are two additional notable changes.
      
      The rule-bending AND+SHIFT optimization in MatchAddress that creates new
      pre-isel nodes during isel is now a little more verbose, but more robust.
      Instead of either creating an invalid DAG or creating an invalid topological
      sort, as it has historically done, it can now just insert the new nodes into
      the node list at a position where they will be consistent with the topological
      ordering.
      
      Also, the address-matching code has logic that checked to see if a node was
      "already selected". However, when a node is selected, it has all its uses
      taken away via ReplaceAllUsesWith or equivalent, so it won't recieve any
      further visits from MatchAddress. This code is now removed.
      
      llvm-svn: 58748
      f14b77eb
    • Evan Cheng's avatar
      Rename isGVLazyPtr to isGVNonLazyPtr relocation. This represents Mac OS X · 132de198
      Evan Cheng authored
      indirect gv reference. Please don't call it lazy.
      
      llvm-svn: 58746
      132de198
  2. Nov 04, 2008
  3. Nov 03, 2008
  4. Oct 31, 2008
  5. Oct 30, 2008
  6. Oct 28, 2008
  7. Oct 27, 2008
    • David Greene's avatar
      · ce2a9381
      David Greene authored
      Have TableGen emit setSubgraphColor calls under control of a -gen-debug
      flag.  Then in a debugger developers can set breakpoints at these calls
      to see waht is about to be selected and what the resulting subgraph
      looks like.  This really helps when debugging instruction selection.
      
      llvm-svn: 58278
      ce2a9381
    • Evan Cheng's avatar
      For now, don't split live intervals around x87 stack register barriers.... · f7137229
      Evan Cheng authored
      For now, don't split live intervals around x87 stack register barriers. FpGET_ST0_80 must be right after a call instruction (and ADJCALLSTACKUP) so we need to find a way to prevent reload of x87 registers between them.
      
      llvm-svn: 58230
      f7137229
  8. Oct 25, 2008
  9. Oct 24, 2008
  10. Oct 22, 2008
  11. Oct 21, 2008
    • Dale Johannesen's avatar
      Add an SSE2 algorithm for uint64->f64 conversion. · 28929589
      Dale Johannesen authored
      The same one Apple gcc uses, faster.  Also gets the
      extreme case in gcc.c-torture/execute/ieee/rbug.c
      correct which we weren't before; this is not
      sufficient to get the test to pass though, there
      is another bug.
      
      llvm-svn: 57926
      28929589
    • Dan Gohman's avatar
      Implement the optimized FCMP_OEQ/FCMP_UNE code for x86 fast-isel. · 4ddf7a4c
      Dan Gohman authored
      llvm-svn: 57915
      4ddf7a4c
    • Dan Gohman's avatar
      Don't create TargetGlobalAddress nodes with offsets that don't fit · 269246b0
      Dan Gohman authored
      in the 32-bit signed offset field of addresses. Even though this
      may be intended, some linkers refuse to relocate code where the
      relocated address computation overflows.
      
      Also, fix the sign-extension of constant offsets to use the
      actual pointer size, rather than the size of the GlobalAddress
      node, which may be different, for example on x86-64 where MVT::i32
      is used when the address is being fit into the 32-bit displacement
      field.
      
      llvm-svn: 57885
      269246b0
    • Dan Gohman's avatar
      Optimized FCMP_OEQ and FCMP_UNE for x86. · 97d95d6d
      Dan Gohman authored
      Where previously LLVM might emit code like this:
      
              ucomisd %xmm1, %xmm0
              setne   %al
              setp    %cl
              orb     %al, %cl
              jne     .LBB4_2
      
      it now emits this:
      
              ucomisd %xmm1, %xmm0
              jne     .LBB4_2
              jp      .LBB4_2
      
      It has fewer instructions and uses fewer registers, but it does
      have more branches. And in the case that this code is followed by
      a non-fallthrough edge, it may be followed by a jmp instruction,
      resulting in three branch instructions in sequence. Some effort
      is made to avoid this situation.
      
      To achieve this, X86ISelLowering.cpp now recognizes FCMP_OEQ and
      FCMP_UNE in lowered form, and replace them with code that emits
      two branches, except in the case where it would require converting
      a fall-through edge to an explicit branch.
      
      Also, X86InstrInfo.cpp's branch analysis and transform code now
      knows now to handle blocks with multiple conditional branches. It
      uses loops instead of having fixed checks for up to two
      instructions. It can now analyze and transform code generated
      from FCMP_OEQ and FCMP_UNE.
      
      llvm-svn: 57873
      97d95d6d
    • Dan Gohman's avatar
      When the coalescer is doing rematerializing, have it remove · c835458d
      Dan Gohman authored
      the copy instruction from the instruction list before asking the
      target to create the new instruction. This gets the old instruction
      out of the way so that it doesn't interfere with the target's
      rematerialization code. In the case of x86, this helps it find
      more cases where EFLAGS is not live.
      
      Also, in the X86InstrInfo.cpp, teach isSafeToClobberEFLAGS to check
      to see if it reached the end of the block after scanning each
      instruction, instead of just before. This lets it notice when the
      end of the block is only two instructions away, without doing any
      additional scanning.
      
      These changes allow rematerialization to clobber EFLAGS in more
      cases, for example using xor instead of mov to set the return value
      to zero in the included testcase.
      
      llvm-svn: 57872
      c835458d
  12. Oct 20, 2008
    • Duncan Sands's avatar
      Have X86 custom lowering for LegalizeTypes use · 1d20ab57
      Duncan Sands authored
      LowerOperation if it doesn't know what else to do.
      This methods should probably be factorized some,
      but this is good enough for the moment.  Have
      LowerATOMIC_BINARY_64 use EXTRACT_ELEMENT rather
      than assuming the operand is a BUILD_PAIR (if it
      is then getNode will automagically simplify the
      EXTRACT_ELEMENT).  This way LowerATOMIC_BINARY_64
      usable from LegalizeTypes.
      
      llvm-svn: 57831
      1d20ab57
  13. Oct 18, 2008
    • Dan Gohman's avatar
      Teach DAGCombine to fold constant offsets into GlobalAddress nodes, · 2fe6bee5
      Dan Gohman authored
      and add a TargetLowering hook for it to use to determine when this
      is legal (i.e. not in PIC mode, etc.)
      
      This allows instruction selection to emit folded constant offsets
      in more cases, such as the included testcase, eliminating the need
      for explicit arithmetic instructions.
      
      This eliminates the need for the C++ code in X86ISelDAGToDAG.cpp
      that attempted to achieve the same effect, but wasn't as effective.
      
      Also, fix handling of offsets in GlobalAddressSDNodes in several
      places, including changing GlobalAddressSDNode's offset from
      int to int64_t.
      
      The Mips, Alpha, Sparc, and CellSPU targets appear to be
      unaware of GlobalAddress offsets currently, so set the hook to
      false on those targets.
      
      llvm-svn: 57748
      2fe6bee5
  14. Oct 17, 2008
Loading