Skip to content
  1. Jul 28, 2013
  2. Jul 22, 2013
  3. Jun 09, 2013
  4. May 21, 2013
  5. Apr 11, 2013
  6. Mar 28, 2013
  7. Mar 26, 2013
  8. Mar 25, 2013
  9. Jan 22, 2013
  10. Nov 21, 2012
  11. Nov 08, 2012
    • Michael Liao's avatar
      Add support of RTM from TSX extension · 73cffddb
      Michael Liao authored
      - Add RTM code generation support throught 3 X86 intrinsics:
        xbegin()/xend() to start/end a transaction region, and xabort() to abort a
        tranaction region
      
      llvm-svn: 167573
      73cffddb
  12. Aug 30, 2012
    • Michael Liao's avatar
      Introduce 'UseSSEx' to force SSE legacy encoding · bbd10792
      Michael Liao authored
      - Add 'UseSSEx' to force SSE legacy insn not being selected when AVX is
        enabled.
      
        As the penalty of inter-mixing SSE and AVX instructions, we need
        prevent SSE legacy insn from being generated except explicitly
        specified through some intrinsics. For patterns supported by both
        SSE and AVX, so far, we force AVX insn will be tried first relying on
        AddedComplexity or position in td file. It's error-prone and
        introduces bugs accidentally.
      
        'UseSSEx' is disabled when AVX is turned on. For SSE insns inherited
        by AVX, we need this predicate to force VEX encoding or SSE legacy
        encoding only.
      
        For insns not inherited by AVX, we still use the previous predicates,
        i.e. 'HasSSEx'. So far, these insns fall into the following
        categories:
        * SSE insns with MMX operands
        * SSE insns with GPR/MEM operands only (xFENCE, PREFETCH, CLFLUSH,
          CRC, and etc.)
        * SSE4A insns.
        * MMX insns.
        * x87 insns added by SSE.
      
      2 test cases are modified:
      
       - test/CodeGen/X86/fast-isel-x86-64.ll
         AVX code generation is different from SSE one. 'vcvtsi2sdq' cannot be
         selected by fast-isel due to complicated pattern and fast-isel
         fallback to materialize it from constant pool.
      
       - test/CodeGen/X86/widen_load-1.ll
         AVX code generation is different from SSE one after fixing SSE/AVX
         inter-mixing. Exec-domain fixing prefers 'vmovapd' instead of
         'vmovaps'.
      
      llvm-svn: 162919
      bbd10792
  13. Jul 30, 2012
  14. Jun 23, 2012
  15. Jun 06, 2012
  16. Jun 03, 2012
  17. May 31, 2012
  18. May 11, 2012
  19. May 02, 2012
  20. May 01, 2012
  21. Feb 19, 2012
  22. Feb 18, 2012
  23. Feb 02, 2012
    • Andrew Trick's avatar
      Instruction scheduling itinerary for Intel Atom. · 8523b16f
      Andrew Trick authored
      Adds an instruction itinerary to all x86 instructions, giving each a default latency of 1, using the InstrItinClass IIC_DEFAULT.
      
      Sets specific latencies for Atom for the instructions in files X86InstrCMovSetCC.td, X86InstrArithmetic.td, X86InstrControl.td, and X86InstrShiftRotate.td. The Atom latencies for the remainder of the x86 instructions will be set in subsequent patches.
      
      Adds a test to verify that the scheduler is working.
      
      Also changes the scheduling preference to "Hybrid" for i386 Atom, while leaving x86_64 as ILP.
      
      Patch by Preston Gurd!
      
      llvm-svn: 149558
      8523b16f
  24. Jan 10, 2012
  25. Jan 09, 2012
  26. Jan 01, 2012
  27. Dec 30, 2011
  28. Dec 29, 2011
  29. Dec 16, 2011
  30. Dec 12, 2011
  31. Dec 09, 2011
  32. Dec 08, 2011
    • Evan Cheng's avatar
      Many of the SSE patterns should not be selected when AVX is available. This... · 4d1a2d44
      Evan Cheng authored
      Many of the SSE patterns should not be selected when AVX is available. This led to the following code in X86Subtarget.cpp
      
        if (HasAVX)
          X86SSELevel = NoMMXSSE;
      
      This is so patterns that are predicated on hasSSE3, etc. would not be selected when avx is available. Instead, the AVX variant is selected.
      However, this breaks instructions which do not have AVX variants.
      
      The right way to fix this is for the SSE but not-AVX patterns to predicate on something like hasSSE3() && !hasAVX().
      Then we can take out the hack in X86Subtarget.cpp. Patterns which do not have AVX variants do not need to change.
      
      However, we need to audit all the patterns before we make the change. This patch is workaround that fixes one specific case,
      the prefetch instructions. rdar://10538297
      
      llvm-svn: 146163
      4d1a2d44
  33. Nov 25, 2011
Loading