Skip to content
  1. Mar 19, 2012
  2. Feb 24, 2012
  3. Feb 16, 2012
  4. Jan 16, 2012
  5. Jan 12, 2012
  6. Dec 24, 2011
    • Chandler Carruth's avatar
      Switch the lowering of CTLZ_ZERO_UNDEF from a .td pattern back to the · 7e9453e9
      Chandler Carruth authored
      X86ISelLowering C++ code. Because this is lowered via an xor wrapped
      around a bsr, we want the dagcombine which runs after isel lowering to
      have a chance to clean things up. In particular, it is very common to
      see code which looks like:
      
        (sizeof(x)*8 - 1) ^ __builtin_clz(x)
      
      Which is trying to compute the most significant bit of 'x'. That's
      actually the value computed directly by the 'bsr' instruction, but if we
      match it too late, we'll get completely redundant xor instructions.
      
      The more naive code for the above (subtracting rather than using an xor)
      still isn't handled correctly due to the dagcombine getting confused.
      
      Also, while here fix an issue spotted by inspection: we should have been
      expanding the zero-undef variants to the normal variants when there is
      an 'lzcnt' instruction. Do so, and test for this. We don't want to
      generate unnecessary 'bsr' instructions.
      
      These two changes fix some regressions in encoding and decoding
      benchmarks. However, there is still a *lot* to be improve on in this
      type of code.
      
      llvm-svn: 147244
      7e9453e9
  7. Dec 20, 2011
    • Chandler Carruth's avatar
      Begin teaching the X86 target how to efficiently codegen patterns that · 24680c24
      Chandler Carruth authored
      use the zero-undefined variants of CTTZ and CTLZ. These are just simple
      patterns for now, there is more to be done to make real world code using
      these constructs be optimized and codegen'ed properly on X86.
      
      The existing tests are spiffed up to check that we no longer generate
      unnecessary cmov instructions, and that we generate the very important
      'xor' to transform bsr which counts the index of the most significant
      one bit to the number of leading (most significant) zero bits. Also they
      now check that when the variant with defined zero result is used, the
      cmov is still produced.
      
      llvm-svn: 146974
      24680c24
  8. Oct 26, 2011
  9. Sep 13, 2011
  10. Sep 07, 2011
  11. Sep 03, 2011
    • Jakob Stoklund Olesen's avatar
      Pseudo CMOV instructions don't clobber EFLAGS. · 1f72dd40
      Jakob Stoklund Olesen authored
      The explanation about a 0 argument being materialized as xor is no
      longer valid.  Rematerialization will check if EFLAGS is live before
      clobbering it.
      
      The code produced by X86TargetLowering::EmitLoweredSelect does not
      clobber EFLAGS.
      
      This causes one less testb instruction to be generated in the cmov.ll
      test case.
      
      llvm-svn: 139057
      1f72dd40
  12. Aug 30, 2011
  13. Aug 26, 2011
  14. Aug 24, 2011
  15. Aug 10, 2011
  16. Jul 27, 2011
  17. Jun 16, 2011
  18. May 21, 2011
  19. May 20, 2011
  20. May 19, 2011
  21. May 17, 2011
  22. May 11, 2011
  23. May 10, 2011
  24. May 08, 2011
    • Benjamin Kramer's avatar
      X86: Add a bunch of peeps for add and sub of SETB. · d724a590
      Benjamin Kramer authored
      "b + ((a < b) ? 1 : 0)" compiles into
      	cmpl	%esi, %edi
      	adcl	$0, %esi
      instead of
      	cmpl	%esi, %edi
      	sbbl	%eax, %eax
      	andl	$1, %eax
      	addl	%esi, %eax
      
      This saves a register, a false dependency on %eax
      (Intel's CPUs still don't ignore it) and it's shorter.
      
      llvm-svn: 131070
      d724a590
  25. Feb 17, 2011
  26. Jan 26, 2011
  27. Jan 18, 2011
  28. Dec 20, 2010
  29. Dec 19, 2010
  30. Dec 15, 2010
  31. Dec 09, 2010
  32. Nov 28, 2010
  33. Nov 27, 2010
  34. Nov 01, 2010
  35. Oct 31, 2010
Loading