Skip to content
  1. May 20, 2013
  2. May 05, 2013
  3. May 01, 2013
    • Filip Pizlo's avatar
      This patch breaks up Wrap.h so that it does not have to include all of · dec20e43
      Filip Pizlo authored
      the things, and renames it to CBindingWrapping.h.  I also moved 
      CBindingWrapping.h into Support/.
      
      This new file just contains the macros for defining different wrap/unwrap 
      methods.
      
      The calls to those macros, as well as any custom wrap/unwrap definitions 
      (like for array of Values for example), are put into corresponding C++ 
      headers.
      
      Doing this required some #include surgery, since some .cpp files relied 
      on the fact that including Wrap.h implicitly caused the inclusion of a 
      bunch of other things.
      
      This also now means that the C++ headers will include their corresponding 
      C API headers; for example Value.h must include llvm-c/Core.h.  I think 
      this is harmless, since the C API headers contain just external function 
      declarations and some C types, so I don't believe there should be any 
      nasty dependency issues here.
      
      llvm-svn: 180881
      dec20e43
    • Richard Trieu's avatar
      Fix a use after free. RI is freed before the call to getDebugLoc(). To · 624c2ebc
      Richard Trieu authored
      prevent this, capture the location before RI is freed.
      
      llvm-svn: 180824
      624c2ebc
  4. Apr 30, 2013
  5. Apr 29, 2013
    • Arnold Schwaighofer's avatar
      SimplifyCFG: If convert single conditional stores · 474df6d3
      Arnold Schwaighofer authored
      This resurrects r179957, but adds code that makes sure we don't touch
      atomic/volatile stores:
      
      This transformation will transform a conditional store with a preceeding
      uncondtional store to the same location:
      
       a[i] =
       may-alias with a[i] load
       if (cond)
         a[i] = Y
      
      into an unconditional store.
      
       a[i] = X
       may-alias with a[i] load
       tmp = cond ? Y : X;
       a[i] = tmp
      
      We assume that on average the cost of a mispredicted branch is going to be
      higher than the cost of a second store to the same location, and that the
      secondary benefits of creating a bigger basic block for other optimizations to
      work on outway the potential case where the branch would be correctly predicted
      and the cost of the executing the second store would be noticably reflected in
      performance.
      
      hmmer's execution time improves by 30% on an imac12,2 on ref data sets. With
      this change we are on par with gcc's performance (gcc also performs this
      transformation). There was a 1.2 % performance improvement on a ARM swift chip.
      Other tests in the test-suite+external seem to be mostly uninfluenced in my
      experiments:
      This optimization was triggered on 41 tests such that the executable was
      different before/after the patch. Only 1 out of the 40 tests (dealII) was
      reproducable below 100% (by about .4%). Given that hmmer benefits so much I
      believe this to be a fair trade off.
      
      llvm-svn: 180731
      474df6d3
  6. Apr 26, 2013
  7. Apr 23, 2013
  8. Apr 21, 2013
    • Arnold Schwaighofer's avatar
      Revert "SimplifyCFG: If convert single conditional stores" · 6eb32b31
      Arnold Schwaighofer authored
      There is the temptation to make this tranform dependent on target information as
      it is not going to be beneficial on all (sub)targets. Therefore, we should
      probably do this in MI Early-Ifconversion.
      
      This reverts commit r179957. Original commit message:
      
      "SimplifyCFG: If convert single conditional stores
      
      This transformation will transform a conditional store with a preceeding
      uncondtional store to the same location:
      
      a[i] =
      may-alias with a[i] load
      if (cond)
          a[i] = Y
      into an unconditional store.
      
      a[i] = X
      may-alias with a[i] load
      tmp = cond ? Y : X;
      a[i] = tmp
      
      We assume that on average the cost of a mispredicted branch is going to be
      higher than the cost of a second store to the same location, and that the
      secondary benefits of creating a bigger basic block for other optimizations to
      work on outway the potential case were the branch would be correctly predicted
      and the cost of the executing the second store would be noticably reflected in
      performance.
      
      hmmer's execution time improves by 30% on an imac12,2 on ref data sets. With
      this change we are on par with gcc's performance (gcc also performs this
      transformation). There was a 1.2 % performance improvement on a ARM swift chip.
      Other tests in the test-suite+external seem to be mostly uninfluenced in my
      experiments:
      This optimization was triggered on 41 tests such that the executable was
      different before/after the patch. Only 1 out of the 40 tests (dealII) was
      reproducable below 100% (by about .4%). Given that hmmer benefits so much I
      believe this to be a fair trade off.
      
      I am going to watch performance numbers across the builtbots and will revert
      this if anything unexpected comes up."
      
      llvm-svn: 179980
      6eb32b31
  9. Apr 20, 2013
    • Arnold Schwaighofer's avatar
      SimplifyCFG: If convert single conditional stores · 3546ccf4
      Arnold Schwaighofer authored
      This transformation will transform a conditional store with a preceeding
      uncondtional store to the same location:
      
       a[i] =
       may-alias with a[i] load
       if (cond)
         a[i] = Y
      
      into an unconditional store.
      
       a[i] = X
       may-alias with a[i] load
       tmp = cond ? Y : X;
       a[i] = tmp
      
      We assume that on average the cost of a mispredicted branch is going to be
      higher than the cost of a second store to the same location, and that the
      secondary benefits of creating a bigger basic block for other optimizations to
      work on outway the potential case were the branch would be correctly predicted
      and the cost of the executing the second store would be noticably reflected in
      performance.
      
      hmmer's execution time improves by 30% on an imac12,2 on ref data sets. With
      this change we are on par with gcc's performance (gcc also performs this
      transformation). There was a 1.2 % performance improvement on a ARM swift chip.
      Other tests in the test-suite+external seem to be mostly uninfluenced in my
      experiments:
      This optimization was triggered on 41 tests such that the executable was
      different before/after the patch. Only 1 out of the 40 tests (dealII) was
      reproducable below 100% (by about .4%). Given that hmmer benefits so much I
      believe this to be a fair trade off.
      
      I am going to watch performance numbers across the builtbots and will revert
      this if anything unexpected comes up.
      
      llvm-svn: 179957
      3546ccf4
  10. Apr 17, 2013
  11. Apr 16, 2013
    • Hans Wennborg's avatar
      simplifycfg: Fix integer overflow converting switch into icmp. · c9e1d992
      Hans Wennborg authored
      If a switch instruction has a case for every possible value of its type,
      with the same successor, SimplifyCFG would replace it with an icmp ult,
      but the computation of the bound overflows in that case, which inverts
      the test.
      
      Patch by Jed Davis!
      
      llvm-svn: 179587
      c9e1d992
  12. Apr 10, 2013
  13. Mar 22, 2013
  14. Mar 12, 2013
    • Meador Inge's avatar
      LibCallSimplifier: optimize speed for short-lived instances · 20255ef2
      Meador Inge authored
      Nadav reported a performance regression due to the work I did to
      merge the library call simplifier into instcombine [1].  The issue
      is that a new LibCallSimplifier object is being created whenever
      InstCombiner::runOnFunction is called.  Every time a LibCallSimplifier
      object is used to optimize a call it creates a hash table to map from
      a function name to an object that optimizes functions of that name.
      For short-lived LibCallSimplifier instances this is quite inefficient.
      Especially for cases where no calls are actually simplified.
      
      This patch fixes the issue by dropping the hash table and implementing
      an explicit lookup function to correlate the function name to the object
      that optimizes functions of that name.  This avoids the cost of always
      building and destroying the hash table in cases where the LibCallSimplifier
      object is short-lived and avoids the cost of building the table when no
      simplifications are actually preformed.
      
      On a benchmark containing 100,000 calls where none of them are simplified
      I noticed a 30% speedup.  On a benchmark containing 100,000 calls where
      all of them are simplified I noticed an 8% speedup.
      
      [1] http://lists.cs.uiuc.edu/pipermail/llvm-commits/Week-of-Mon-20130304/167639.html
      
      llvm-svn: 176840
      20255ef2
  15. Mar 11, 2013
  16. Mar 07, 2013
    • Pekka Jaaskelainen's avatar
      Fixed a crash when cloning a function into a function with · 093cf41e
      Pekka Jaaskelainen authored
      different size argument list and without attributes in the
      arguments.
      
      llvm-svn: 176632
      093cf41e
    • Andrew Trick's avatar
      SimplifyCFG fix for volatile load/store. · a0a5ca06
      Andrew Trick authored
      Fixes rdar:13349374.
      
      Volatile loads and stores need to be preserved even if the language
      standard says they are undefined. "volatile" in this context means "get
      out of the way compiler, let my platform handle it".
      
      Additionally, this is the only way I know of with llvm to write to the
      first page (when hardware allows) without dropping to assembly.
      
      llvm-svn: 176599
      a0a5ca06
  17. Mar 04, 2013
    • Preston Gurd's avatar
      Bypass Slow Divides · 485296d1
      Preston Gurd authored
      * Only apply divide bypass optimization when not optimizing for size. 
      * Fixed bug caused by constant for 0 value of type Int32,
        used dividend type to generate the constant instead.
      * For atom x86-64 apply the divide bypass to use 16-bit divides instead of
        64-bit divides when operand values are small enough.
      * Added lit tests for 64-bit divide bypass.
      
      Patch by Tyler Nowicki!
      
      llvm-svn: 176442
      485296d1
  18. Mar 02, 2013
  19. Feb 27, 2013
  20. Feb 22, 2013
    • Bill Wendling's avatar
      Implement the NoBuiltin attribute. · 09bd1f71
      Bill Wendling authored
      The 'nobuiltin' attribute is applied to call sites to indicate that LLVM should
      not treat the callee function as a built-in function. I.e., it shouldn't try to
      replace that function with different code.
      
      llvm-svn: 175835
      09bd1f71
  21. Feb 19, 2013
  22. Feb 09, 2013
  23. Feb 08, 2013
  24. Feb 05, 2013
  25. Jan 31, 2013
  26. Jan 30, 2013
    • Manman Ren's avatar
      Linker: correctly link in dbg.declare · 81dcc628
      Manman Ren authored
      Given source IR:
      call void @llvm.dbg.declare(metadata !{i32* %argc.addr}, metadata !14), !dbg !15
      we used to generate 
      call void @llvm.dbg.declare(metadata !27, metadata !28), !dbg !29
      !27 = metadata !{null}
      
      With this patch, we will correctly generate
      call void @llvm.dbg.declare(metadata !{i32* %argc.addr}, metadata !27), !dbg !28
      
      Looking up %argc.addr in ValueMap will return null, since %argc.addr is already
      correctly set up, we can use identity mapping.
      
      llvm-svn: 173946
      81dcc628
  27. Jan 27, 2013
Loading