Remove the use of TargetMachine from X86InstrInfo.
[oota-llvm.git] / lib / Target / R600 / SILowerI1Copies.cpp
1 //===-- SILowerI1Copies.cpp - Lower I1 Copies -----------------------------===//
2 //
3 //                     The LLVM Compiler Infrastructure
4 //
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
7 //
8 /// i1 values are usually inserted by the CFG Structurize pass and they are
9 /// unique in that they can be copied from VALU to SALU registers.
10 /// This is not possible for any other value type.  Since there are no
11 /// MOV instructions for i1, we to use V_CMP_* and V_CNDMASK to move the i1.
12 ///
13 //===----------------------------------------------------------------------===//
14 //
15
16 #define DEBUG_TYPE "si-i1-copies"
17 #include "AMDGPU.h"
18 #include "SIInstrInfo.h"
19 #include "llvm/CodeGen/LiveIntervalAnalysis.h"
20 #include "llvm/CodeGen/MachineDominators.h"
21 #include "llvm/CodeGen/MachineFunctionPass.h"
22 #include "llvm/CodeGen/MachineInstrBuilder.h"
23 #include "llvm/CodeGen/MachineRegisterInfo.h"
24 #include "llvm/IR/LLVMContext.h"
25 #include "llvm/IR/Function.h"
26 #include "llvm/Support/Debug.h"
27 #include "llvm/Target/TargetMachine.h"
28
29 using namespace llvm;
30
31 namespace {
32
33 class SILowerI1Copies : public MachineFunctionPass {
34 public:
35   static char ID;
36
37 public:
38   SILowerI1Copies() : MachineFunctionPass(ID) {
39     initializeSILowerI1CopiesPass(*PassRegistry::getPassRegistry());
40   }
41
42   virtual bool runOnMachineFunction(MachineFunction &MF) override;
43
44   virtual const char *getPassName() const override {
45     return "SI Lower il Copies";
46   }
47
48   virtual void getAnalysisUsage(AnalysisUsage &AU) const override {
49   AU.addRequired<MachineDominatorTree>();
50     AU.setPreservesCFG();
51     MachineFunctionPass::getAnalysisUsage(AU);
52   }
53 };
54
55 } // End anonymous namespace.
56
57 INITIALIZE_PASS_BEGIN(SILowerI1Copies, DEBUG_TYPE,
58                       "SI Lower il Copies", false, false)
59 INITIALIZE_PASS_DEPENDENCY(MachineDominatorTree)
60 INITIALIZE_PASS_END(SILowerI1Copies, DEBUG_TYPE,
61                     "SI Lower il Copies", false, false)
62
63 char SILowerI1Copies::ID = 0;
64
65 char &llvm::SILowerI1CopiesID = SILowerI1Copies::ID;
66
67 FunctionPass *llvm::createSILowerI1CopiesPass() {
68   return new SILowerI1Copies();
69 }
70
71 bool SILowerI1Copies::runOnMachineFunction(MachineFunction &MF) {
72   MachineRegisterInfo &MRI = MF.getRegInfo();
73   const SIInstrInfo *TII = static_cast<const SIInstrInfo *>(
74       MF.getTarget().getInstrInfo());
75   const TargetRegisterInfo *TRI = MF.getTarget().getRegisterInfo();
76   std::vector<unsigned> I1Defs;
77
78   for (MachineFunction::iterator BI = MF.begin(), BE = MF.end();
79                                                   BI != BE; ++BI) {
80
81     MachineBasicBlock &MBB = *BI;
82     MachineBasicBlock::iterator I, Next;
83     for (I = MBB.begin(); I != MBB.end(); I = Next) {
84       Next = std::next(I);
85       MachineInstr &MI = *I;
86
87       if (MI.getOpcode() == AMDGPU::V_MOV_I1) {
88         I1Defs.push_back(MI.getOperand(0).getReg());
89         MI.setDesc(TII->get(AMDGPU::V_MOV_B32_e32));
90         continue;
91       }
92
93       if (MI.getOpcode() == AMDGPU::V_AND_I1) {
94         I1Defs.push_back(MI.getOperand(0).getReg());
95         MI.setDesc(TII->get(AMDGPU::V_AND_B32_e32));
96         continue;
97       }
98
99       if (MI.getOpcode() == AMDGPU::V_OR_I1) {
100         I1Defs.push_back(MI.getOperand(0).getReg());
101         MI.setDesc(TII->get(AMDGPU::V_OR_B32_e32));
102         continue;
103       }
104
105       if (MI.getOpcode() != AMDGPU::COPY ||
106           !TargetRegisterInfo::isVirtualRegister(MI.getOperand(0).getReg()) ||
107           !TargetRegisterInfo::isVirtualRegister(MI.getOperand(1).getReg()))
108         continue;
109
110
111       const TargetRegisterClass *DstRC =
112           MRI.getRegClass(MI.getOperand(0).getReg());
113       const TargetRegisterClass *SrcRC =
114           MRI.getRegClass(MI.getOperand(1).getReg());
115
116       if (DstRC == &AMDGPU::VReg_1RegClass &&
117           TRI->getCommonSubClass(SrcRC, &AMDGPU::SGPR_64RegClass)) {
118         I1Defs.push_back(MI.getOperand(0).getReg());
119         BuildMI(MBB, &MI, MI.getDebugLoc(), TII->get(AMDGPU::V_CNDMASK_B32_e64))
120                 .addOperand(MI.getOperand(0))
121                 .addImm(0)
122                 .addImm(-1)
123                 .addOperand(MI.getOperand(1))
124                 .addImm(0)
125                 .addImm(0)
126                 .addImm(0)
127                 .addImm(0);
128         MI.eraseFromParent();
129       } else if (TRI->getCommonSubClass(DstRC, &AMDGPU::SGPR_64RegClass) &&
130                  SrcRC == &AMDGPU::VReg_1RegClass) {
131         BuildMI(MBB, &MI, MI.getDebugLoc(), TII->get(AMDGPU::V_CMP_NE_I32_e64))
132                 .addOperand(MI.getOperand(0))
133                 .addImm(0)
134                 .addOperand(MI.getOperand(1))
135                 .addImm(0)
136                 .addImm(0)
137                 .addImm(0)
138                 .addImm(0);
139         MI.eraseFromParent();
140       }
141     }
142   }
143
144   for (unsigned Reg : I1Defs)
145     MRI.setRegClass(Reg, &AMDGPU::VReg_32RegClass);
146
147   return false;
148 }