Skip to content
  1. Jun 12, 2015
    • John Brawn's avatar
      [ARM] Disabling vfp4 should disable fp16 · d9e39d53
      John Brawn authored
      ARMTargetParser::getFPUFeatures should disable fp16 whenever it
      disables vfp4, as otherwise something like -mcpu=cortex-a7 -mfpu=none
      leaves us with fp16 enabled (though the only effect that will have is
      a wrong build attribute).
      
      Differential Revision: http://reviews.llvm.org/D10397
      
      llvm-svn: 239599
      d9e39d53
    • Peter Collingbourne's avatar
      LowerBitSets: Give names to aliases of unnamed bitset element objects. · 005354b1
      Peter Collingbourne authored
      It is valid for globals to be unnamed, but aliases must have a name. To avoid
      creating invalid IR, we need to assign names to any aliases we create that
      point to unnamed objects that have been moved into combined globals.
      
      llvm-svn: 239590
      005354b1
    • Alexey Samsonov's avatar
      [GVN] Use a simpler form of IRBuilder constructor. · 9947e48c
      Alexey Samsonov authored
      Summary:
      A side effect of this change is that it IRBuilder now automatically
      created debug info locations for new instructions, which is the
      same as debug location of insertion point. This is fine for the
      functions in questions (GetStoreValueForLoad and
      GetMemInstValueForLoad), as they are used in two situations:
        * GVN::processLoad, which tries to eliminate a load. In this case
          new instructions would have the same debug location as the load they
          eventually replace;
        * MaterializeAdjustedValue, which adds new instructions to the end
          of the basic blocks, which could later be used to replace the load
          definition. In this case we don't yet know the way the load would
          be eventually replaced (either by assembling the precomputed values
          via PHI, or by using them directly), so just using the basic block
          strategy seems to be reasonable. There is also a special case
          in the code that *would* adjust the location of the last
          instruction replacing the load definition to the location of the
          load.
      
      Test Plan: regression test suite
      
      Reviewers: echristo, dberlin, dblaikie
      
      Subscribers: llvm-commits
      
      Differential Revision: http://reviews.llvm.org/D10405
      
      llvm-svn: 239585
      9947e48c
    • Reid Kleckner's avatar
      [WinEH] Put finally pointers in the handler scope table field · 81d1cc00
      Reid Kleckner authored
      We were putting them in the filter field, which is correct for 64-bit
      but wrong for 32-bit.
      
      Also switch the order of scope table entry emission so outermost entries
      are emitted first, and fix an obvious state assignment bug.
      
      llvm-svn: 239574
      81d1cc00
    • Reid Kleckner's avatar
      [WinEH] Create an llvm.x86.seh.exceptioninfo intrinsic · a9d62535
      Reid Kleckner authored
      This intrinsic is like framerecover plus a load. It recovers the EH
      registration stack allocation from the parent frame and loads the
      exception information field out of it, giving back a pointer to an
      EXCEPTION_POINTERS struct. It's designed for clang to use in SEH filter
      expressions instead of accessing the EXCEPTION_POINTERS parameter that
      is available on x64.
      
      This required a minor change to MC to allow defining a label variable to
      another absolute framerecover label variable.
      
      llvm-svn: 239567
      a9d62535
  2. Jun 11, 2015
    • Peter Collingbourne's avatar
      Object: Prepend __imp_ when mangling a dllimport symbol in IRObjectFile. · 82e657b5
      Peter Collingbourne authored
      We cannot prepend __imp_ in the IR mangler because a function reference may
      be emitted unmangled in a constant initializer. The linker is expected to
      resolve such references to thunks. This is covered by the new test case.
      
      Strictly speaking we ought to emit two undefined symbols, one with __imp_ and
      one without, as we cannot know which symbol the final object file will refer
      to. However, this would require rather intrusive changes to IRObjectFile,
      and lld works fine without it for now.
      
      This reimplements r239437, which was reverted in r239502.
      
      Differential Revision: http://reviews.llvm.org/D10400
      
      llvm-svn: 239560
      82e657b5
    • Alexey Samsonov's avatar
      Set proper debug location for branch added in BasicBlock::splitBasicBlock(). · 770f65ca
      Alexey Samsonov authored
      This improves debug locations in passes that do a lot of basic block
      transformations. Important case is LoopUnroll pass, the test for correct
      debug locations accompanies this change.
      
      Test Plan: regression test suite
      
      Reviewers: dblaikie, sanjoy
      
      Subscribers: llvm-commits
      
      Differential Revision: http://reviews.llvm.org/D10367
      
      llvm-svn: 239551
      770f65ca
    • Rafael Espindola's avatar
      This reverts commit r239529 and r239514. · 65d37e64
      Rafael Espindola authored
      Revert "[AArch64] Match interleaved memory accesses into ldN/stN instructions."
      Revert "Fixing MSVC 2013 build error."
      
      The  test/CodeGen/AArch64/aarch64-interleaved-accesses.ll test was failing on OS X.
      
      llvm-svn: 239544
      65d37e64
    • Reid Kleckner's avatar
      Revert "Fix merges of non-zero vector stores" · 2691c59e
      Reid Kleckner authored
      This reverts commit r239539.
      
      It was causing SDAG assertions while building freetype.
      
      llvm-svn: 239543
      2691c59e
    • Matt Arsenault's avatar
      SLSR: Pass address space to isLegalAddressingMode · 91f90e69
      Matt Arsenault authored
      This only updates one of the uses. The other is used in cases
      that may never touch memory, so I'm not sure why this is even
      calling it. Perhaps there should be a new, similar hook for such
      cases or pass -1 for unknown address space.
      
      llvm-svn: 239540
      91f90e69
    • Matt Arsenault's avatar
      Fix merges of non-zero vector stores · e23a063d
      Matt Arsenault authored
      Now actually stores the non-zero constant instead of 0.
      I somehow forgot to include this part of r238108.
      
      The test change was just an independent instruction order swap,
      so just add another check line to satisfy CHECK-NEXT.
      
      llvm-svn: 239539
      e23a063d
    • Tom Stellard's avatar
      R600/SI: Add -mcpu=bonaire to a test that uses flat address space · 53e015f3
      Tom Stellard authored
      Flat instructions don't exist on SI, but there is a bug in the backend that
      allows them to be selected.
      
      llvm-svn: 239533
      53e015f3
    • Toma Tabacu's avatar
      Recommit "[mips] [IAS] Add support for BNE and BEQ with an immediate operand." (r239396). · e1e460db
      Toma Tabacu authored
      Apparently, Arcanist didn't include some of my local changes in my previous
      commit attempt.
      
      llvm-svn: 239523
      e1e460db
    • Zoran Jovanovic's avatar
    • Zoran Jovanovic's avatar
    • Zoran Jovanovic's avatar
      [mips][microMIPSr6] Change disassembler tests to one line format · fcecf260
      Zoran Jovanovic authored
      llvm-svn: 239519
      fcecf260
    • Hao Liu's avatar
      [AArch64] Match interleaved memory accesses into ldN/stN instructions. · 4566d18e
      Hao Liu authored
      Add a pass AArch64InterleavedAccess to identify and match interleaved memory accesses. This pass transforms an interleaved load/store into ldN/stN intrinsic. As Loop Vectorizor disables optimization on interleaved accesses by default, this optimization is also disabled by default. To enable it by "-aarch64-interleaved-access-opt=true"
      
      E.g. Transform an interleaved load (Factor = 2):
             %wide.vec = load <8 x i32>, <8 x i32>* %ptr
             %v0 = shuffle %wide.vec, undef, <0, 2, 4, 6>  ; Extract even elements
             %v1 = shuffle %wide.vec, undef, <1, 3, 5, 7>  ; Extract odd elements
           Into:
             %ld2 = { <4 x i32>, <4 x i32> } call aarch64.neon.ld2(%ptr)
             %v0 = extractelement { <4 x i32>, <4 x i32> } %ld2, i32 0
             %v1 = extractelement { <4 x i32>, <4 x i32> } %ld2, i32 1
      
      E.g. Transform an interleaved store (Factor = 2):
             %i.vec = shuffle %v0, %v1, <0, 4, 1, 5, 2, 6, 3, 7>  ; Interleaved vec
             store <8 x i32> %i.vec, <8 x i32>* %ptr
           Into:
             %v0 = shuffle %i.vec, undef, <0, 1, 2, 3>
             %v1 = shuffle %i.vec, undef, <4, 5, 6, 7>
             call void aarch64.neon.st2(%v0, %v1, %ptr)
      
      llvm-svn: 239514
      4566d18e
    • Simon Pilgrim's avatar
      [X86][SSE] Vectorized i8 and i16 shift operators · 5965680d
      Simon Pilgrim authored
      This patch ensures that SHL/SRL/SRA shifts for i8 and i16 vectors avoid scalarization. It builds on the existing i8 SHL vectorized implementation of moving the shift bits up to the sign bit position and separating the 4, 2 & 1 bit shifts with several improvements:
      
      1 - SSE41 targets can use (v)pblendvb directly with the sign bit instead of performing a comparison to feed into a VSELECT node.
      2 - pre-SSE41 targets were masking + comparing with an 0x80 constant - we avoid this by using the fact that a set sign bit means a negative integer which can be compared against zero to then feed into VSELECT, avoiding the need for a constant mask (zero generation is much cheaper).
      3 - SRA i8 needs to be unpacked to the upper byte of a i16 so that the i16 psraw instruction can be correctly used for sign extension - we have to do more work than for SHL/SRL but perf tests indicate that this is still beneficial.
      
      The i16 implementation is similar but simpler than for i8 - we have to do 8, 4, 2 & 1 bit shifts but less shift masking is involved. SSE41 use of (v)pblendvb requires that the i16 shift amount is splatted to both bytes however.
      
      Tested on SSE2, SSE41 and AVX machines.
      
      Differential Revision: http://reviews.llvm.org/D9474
      
      llvm-svn: 239509
      5965680d
    • Nemanja Ivanovic's avatar
      LLVM support for vector quad bit permute and gather instructions through builtins · ea1db8a6
      Nemanja Ivanovic authored
      This patch corresponds to review:
      http://reviews.llvm.org/D10096
      
      This is the back end portion of the patch related to D10095.
      The patch adds the instructions and back end intrinsics for:
      vbpermq
      vgbbd
      
      llvm-svn: 239505
      ea1db8a6
    • Reid Kleckner's avatar
      Revert "Move dllimport name mangling to IR mangler." · c35e7f52
      Reid Kleckner authored
      This reverts commit r239437.
      
      This broke clang-cl self-hosts. We'd end up calling the __imp_ symbol
      directly instead of using it to do an indirect function call.
      
      llvm-svn: 239502
      c35e7f52
  3. Jun 10, 2015
  4. Jun 09, 2015
Loading