Skip to content
  1. Jul 24, 2018
  2. Jul 23, 2018
  3. Jul 21, 2018
    • Zachary Turner's avatar
      [llvm-undname] Flush output before demangling. · c93b870c
      Zachary Turner authored
      If an error occurs and we write it to stderr, it could appear
      before we wrote the mangled name which we're undecorating.
      By flushing stdout first, we ensure that the messages are always
      sequenced in the correct order.
      
      llvm-svn: 337645
      c93b870c
  4. Jul 20, 2018
    • Martin Storsjö's avatar
      [llvm-undname] Remove a superfluous semicolon. NFC. · 21524bed
      Martin Storsjö authored
      llvm-svn: 337615
      21524bed
    • Jordan Rupprecht's avatar
      [llvm-objcopy] Add basic support for --rename-section · db2036e1
      Jordan Rupprecht authored
      Summary:
      Add basic support for --rename-section=old=new to llvm-objcopy.
      
      A full replacement for GNU objcopy requires also modifying flags (i.e. --rename-section=old=new,flag1,flag2); I'd like to keep that in a separate change to keep this simple.
      
      Reviewers: jakehehrlich, alexshap
      
      Subscribers: llvm-commits
      
      Differential Revision: https://reviews.llvm.org/D49576
      
      llvm-svn: 337604
      db2036e1
    • Zachary Turner's avatar
      Add a Microsoft Demangler. · f435a7ea
      Zachary Turner authored
      This adds initial support for a demangling library (LLVMDemangle)
      and tool (llvm-undname) for demangling Microsoft names.  This
      doesn't cover 100% of cases and there are some known limitations
      which I intend to address in followup patches, at least until such
      time that we have (near) 100% test coverage matching up with all
      of the test cases in clang/test/CodeGenCXX/mangle-ms-*.
      
      Differential Revision: https://reviews.llvm.org/D49552
      
      llvm-svn: 337584
      f435a7ea
    • Zachary Turner's avatar
      Rewrite the VS integration scripts. · 83226b91
      Zachary Turner authored
      This is a new modernized VS integration installer.  It adds a
      Visual Studio .sln file which, when built, outputs a VSIX that can
      be used to install ourselves as a "real" Visual Studio Extension.
      We can even upload this extension to the visual studio marketplace.
      
      This fixes a longstanding problem where we didn't support installing
      into VS 2017 and higher.  In addition to supporting VS 2017, due
      to the way this is written we now longer need to do anything special
      to support future versions of VS as well.  Everything should
      "just work".  This also fixes several bugs with our old integration,
      such as MSBuild triggering full rebuilds when /Zi was used.
      
      Finally, we add a new UI page called "LLVM" which becomes visible
      when the LLVM toolchain is selected.  For now this only contains
      one option which is the path to clang-cl.exe, but in the future
      we can add more things here.
      
      Differential Revision: https://reviews.llvm.org/D42762
      
      llvm-svn: 337572
      83226b91
  5. Jul 19, 2018
  6. Jul 18, 2018
  7. Jul 17, 2018
  8. Jul 16, 2018
    • Jake Ehrlich's avatar
      [llvm-objcopy] Add support for large indexes · c7f8ac78
      Jake Ehrlich authored
      This patch is an update of an older patch that never landed
      (see here: https://reviews.llvm.org/D42516)
      
      Recently various users have run into this issue and it just 100%
      has to be solved at this point. The main difference in this patch
      is that I use gunzip instead of unzip which should hopefully allow
      tests to pass. Please review this as if it is a new patch however.
      I found some issues along the way and made some minor modifications.
      
      The binary used in this patch for testing (a zip file to make it small)
      can be found here:
      https://drive.google.com/file/d/1UjsnTO9edLttZibbr-2T1bJl92KEQFAO/view?usp=sharing
      
      Differential Revision: https://reviews.llvm.org/D49206
      
      llvm-svn: 337204
      c7f8ac78
    • Teresa Johnson's avatar
      Restore "[ThinLTO] Ensure we always select the same function copy to import" · d68935c5
      Teresa Johnson authored
      This reverts commit r337081, therefore restoring r337050 (and fix in
      r337059), with test fix for bot failure described after the original
      description below.
      
      In order to always import the same copy of a linkonce function,
      even when encountering it with different thresholds (a higher one then a
      lower one), keep track of the summary we decided to import.
      This ensures that the backend only gets a single definition to import
      for each GUID, so that it doesn't need to choose one.
      
      Move the largest threshold the GUID was considered for import into the
      current module out of the ImportMap (which is part of a larger map
      maintained across the whole index), and into a new map just maintained
      for the current module we are computing imports for. This saves some
      memory since we no longer have the thresholds maintained across the
      whole index (and throughout the in-process backends when doing a normal
      non-distributed ThinLTO build), at the cost of some additional
      information being maintained for each invocation of ComputeImportForModule
      (the selected summary pointer for each import).
      
      There is an additional map lookup for each callee being considered for
      importing, however, this was able to subsume a map lookup in the
      Worklist iteration that invokes computeImportForFunction. We also are
      able to avoid calling selectCallee if we already failed to import at the
      same or higher threshold.
      
      I compared the run time and peak memory for the SPEC2006 471.omnetpp
      benchmark (running in-process ThinLTO backends), as well as for a large
      internal benchmark with a distributed ThinLTO build (so just looking at
      the thin link time/memory). Across a number of runs with and without
      this change there was no significant change in the time and memory.
      
      (I tried a few other variations of the change but they also didn't
      improve time or peak memory).
      
      The new commit removes a test that no longer makes sense
      (Transforms/FunctionImport/hotness_based_import2.ll), as exposed by the
      reverse-iteration bot. The test depends on the order of processing the
      summary call edges, and actually depended on the old problematic
      behavior of selecting more than one summary for a given GUID when
      encountered with different thresholds. There was no guarantee even
      before that we would eventually pick the linkonce copy with the hottest
      call edges, it just happened to work with the test and the old code, and
      there was no guarantee that we would end up importing the selected
      version of the copy that had the hottest call edges (since the backend
      would effectively import only one of the selected copies).
      
      Reviewers: davidxl
      
      Subscribers: mehdi_amini, inglorion, llvm-commits
      
      Differential Revision: https://reviews.llvm.org/D48670
      
      llvm-svn: 337184
      d68935c5
    • Joel Galenson's avatar
      [cfi-verify] Abort on unsupported targets · 4099b249
      Joel Galenson authored
      As suggested in the review for r337007, this makes cfi-verify abort on unsupported targets instead of producing incorrect results.  It also updates the design document to reflect this.
      
      Differential Revision: https://reviews.llvm.org/D49304
      
      llvm-svn: 337181
      4099b249
  9. Jul 15, 2018
    • Andrea Di Biagio's avatar
      [llvm-mca][BtVer2] teach how to identify false dependencies on partially written · ff630c2c
      Andrea Di Biagio authored
      registers.
      
      The goal of this patch is to improve the throughput analysis in llvm-mca for the
      case where instructions perform partial register writes.
      
      On x86, partial register writes are quite difficult to model, mainly because
      different processors tend to implement different register merging schemes in
      hardware.
      
      When the code contains partial register writes, the IPC (instructions per
      cycles) estimated by llvm-mca tends to diverge quite significantly from the
      observed IPC (using perf).
      
      Modern AMD processors (at least, from Bulldozer onwards) don't rename partial
      registers. Quoting Agner Fog's microarchitecture.pdf:
      " The processor always keeps the different parts of an integer register together.
      For example, AL and AH are not treated as independent by the out-of-order
      execution mechanism. An instruction that writes to part of a register will
      therefore have a false dependence on any previous write to the same register or
      any part of it."
      
      This patch is a first important step towards improving the analysis of partial
      register updates. It changes the semantic of RegisterFile descriptors in
      tablegen, and teaches llvm-mca how to identify false dependences in the presence
      of partial register writes (for more details: see the new code comments in
      include/Target/TargetSchedule.h - class RegisterFile).
      
      This patch doesn't address the case where a write to a part of a register is
      followed by a read from the whole register.  On Intel chips, high8 registers
      (AH/BH/CH/DH)) can be stored in separate physical registers. However, a later
      (dirty) read of the full register (example: AX/EAX) triggers a merge uOp, which
      adds extra latency (and potentially affects the pipe usage).
      This is a very interesting article on the subject with a very informative answer
      from Peter Cordes:
      https://stackoverflow.com/questions/45660139/how-exactly-do-partial-registers-on-haswell-skylake-perform-writing-al-seems-to
      
      In future, the definition of RegisterFile can be extended with extra information
      that may be used to identify delays caused by merge opcodes triggered by a dirty
      read of a partial write.
      
      Differential Revision: https://reviews.llvm.org/D49196
      
      llvm-svn: 337123
      ff630c2c
    • Matt Davis's avatar
      [llvm-mca] Turn InstructionTables into a Stage. · 0e8402ec
      Matt Davis authored
      Summary:
      This patch converts the InstructionTables class into a subclass of mca::Stage.  This change allows us to use the Stage's inherited Listeners for event notifications.  This also allows us to create a simple pipeline for viewing the InstructionTables report.
      
      I have been working on a follow on patch that should cleanup addView in InstructionTables.  Right now, addView adds the view to both the Listener list and Views list.  The follow-on patch addresses the fact that we don't really need two lists in this case.  That change is not specific to just InstructionTables, so it will be a separate patch. 
      
      Reviewers: andreadb, courbet, RKSimon
      
      Reviewed By: andreadb
      
      Subscribers: tschuett, gbedwell, llvm-commits
      
      Differential Revision: https://reviews.llvm.org/D49329
      
      llvm-svn: 337113
      0e8402ec
  10. Jul 14, 2018
  11. Jul 13, 2018
    • Teresa Johnson's avatar
      [ThinLTO] Ensure we always select the same function copy to import · d94c0594
      Teresa Johnson authored
      In order to always import the same copy of a linkonce function,
      even when encountering it with different thresholds (a higher one then a
      lower one), keep track of the summary we decided to import.
      This ensures that the backend only gets a single definition to import
      for each GUID, so that it doesn't need to choose one.
      
      Move the largest threshold the GUID was considered for import into the
      current module out of the ImportMap (which is part of a larger map
      maintained across the whole index), and into a new map just maintained
      for the current module we are computing imports for. This saves some
      memory since we no longer have the thresholds maintained across the
      whole index (and throughout the in-process backends when doing a normal
      non-distributed ThinLTO build), at the cost of some additional
      information being maintained for each invocation of ComputeImportForModule
      (the selected summary pointer for each import).
      
      There is an additional map lookup for each callee being considered for
      importing, however, this was able to subsume a map lookup in the
      Worklist iteration that invokes computeImportForFunction. We also are
      able to avoid calling selectCallee if we already failed to import at the
      same or higher threshold.
      
      I compared the run time and peak memory for the SPEC2006 471.omnetpp
      benchmark (running in-process ThinLTO backends), as well as for a large
      internal benchmark with a distributed ThinLTO build (so just looking at
      the thin link time/memory). Across a number of runs with and without
      this change there was no significant change in the time and memory.
      
      (I tried a few other variations of the change but they also didn't
      improve time or peak memory).
      
      Reviewers: davidxl
      
      Subscribers: mehdi_amini, inglorion, llvm-commits
      
      Differential Revision: https://reviews.llvm.org/D48670
      
      llvm-svn: 337050
      d94c0594
    • Joel Galenson's avatar
      [cfi-verify] Support AArch64. · 06e7e579
      Joel Galenson authored
      This patch adds support for AArch64 to cfi-verify.
      
      This required three changes to cfi-verify.  First, it generalizes checking if an instruction is a trap by adding a new isTrap flag to TableGen (and defining it for x86 and AArch64).  Second, the code that ensures that the operand register is not clobbered between the CFI check and the indirect call needs to allow a single dereference (in x86 this happens as part of the jump instruction).  Third, we needed to ensure that return instructions are not counted as indirect branches.  Technically, returns are indirect branches and can be covered by CFI, but LLVM's forward-edge CFI does not protect them, and x86 does not consider them, so we keep that behavior.
      
      In addition, we had to improve AArch64's code to evaluate the branch target of a MCInst to handle calls where the destination is not the first operand (which it often is not).
      
      Differential Revision: https://reviews.llvm.org/D48836
      
      llvm-svn: 337007
      06e7e579
    • Andrea Di Biagio's avatar
      [llvm-mca] Improve a few debug prints. NFC · 23fbe7cb
      Andrea Di Biagio authored
      llvm-svn: 337003
      23fbe7cb
    • Andrea Di Biagio's avatar
      [llvm-mca] Simplify the Pipeline constructor. NFC · 4335b3e2
      Andrea Di Biagio authored
      llvm-svn: 336984
      4335b3e2
    • Andrea Di Biagio's avatar
      ef6b8a33
    • Dean Michael Berris's avatar
      [XRay][compiler-rt] Add PID field to llvm-xray tool and add PID metadata record entry in FDR mode · 10141261
      Dean Michael Berris authored
      Summary:
      llvm-xray changes:
      - account-mode - process-id  {...} shows after thread-id
      - convert-mode - process {...} shows after thread
      - parses FDR and basic mode pid entries
      - Checks version number for FDR log parsing.
      
      Basic logging changes:
      - Update header version from 2 -> 3
      
      FDR logging changes:
      - Update header version from 2 -> 3
      - in writeBufferPreamble, there is an additional PID Metadata record (after thread id record and tsc record)
      
      Test cases changes:
      - fdr-mode.cc, fdr-single-thread.cc, fdr-thread-order.cc modified to catch process id output in the log.
      
      Reviewers: dberris
      
      Reviewed By: dberris
      
      Subscribers: hiraditya, llvm-commits, #sanitizers
      
      Differential Revision: https://reviews.llvm.org/D49153
      
      llvm-svn: 336974
      10141261
    • Matt Davis's avatar
      [llvm-mca] Constify SourceMgr::hasNext. NFC. · 73dd5f9c
      Matt Davis authored
      llvm-svn: 336961
      73dd5f9c
    • Matt Davis's avatar
      [llvm-mca] Add cycleBegin/cycleEnd callbacks to mca::Stage. · 32508996
      Matt Davis authored
      Summary:
      This patch  clears up some of the semantics within the Stage class.  Now, preExecute
      can be called multiple times per simulated cycle.  Previously preExecute was
      only called once per cycle, and postExecute could have been called multiple
      times.
      
      Now, cycleStart/cycleEnd are called only once per simulated cycle.
      preExecute/postExecute can be called multiple times per cycle.  This
      occurs because multiple execution events can occur during a single cycle.
      
      When stages are executed (Pipeline::runCycle), the postExecute hook will
      be called only if all Stages return a success from their 'execute' callback.
      
      Reviewers: andreadb, courbet, RKSimon
      
      Reviewed By: andreadb
      
      Subscribers: tschuett, gbedwell, llvm-commits
      
      Differential Revision: https://reviews.llvm.org/D49250
      
      llvm-svn: 336959
      32508996
  12. Jul 12, 2018
  13. Jul 11, 2018
Loading