| /netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
| H A D | MachineTraceMetrics.cpp | 629 unsigned DefOp; member 632 DataDep(const MachineInstr *DefMI, unsigned DefOp, unsigned UseOp) in DataDep() 633 : DefMI(DefMI), DefOp(DefOp), UseOp(UseOp) {} in DataDep() 642 DefOp = DefI.getOperandNo(); in DataDep() 740 for (unsigned DefOp : LiveDefOps) { in updatePhysDepsDownwards() local 741 for (MCRegUnitIterator Units(UseMI->getOperand(DefOp).getReg().asMCReg(), in updatePhysDepsDownwards() 746 LRU.Op = DefOp; in updatePhysDepsDownwards() 804 .computeOperandLatency(Dep.DefMI, Dep.DefOp, &UseMI, Dep.UseOp); in updateDepth() 959 UseHeight += SchedModel.computeOperandLatency(Dep.DefMI, Dep.DefOp, &UseMI, in pushDepHeight() 979 addLiveIns(const MachineInstr *DefMI, unsigned DefOp, in addLiveIns() argument [all …]
|
| H A D | PeepholeOptimizer.cpp | 1519 MachineOperand &DefOp = MI.getOperand(0); in findTargetRecurrence() local 1520 if (!isVirtualRegisterOperand(DefOp)) in findTargetRecurrence() 1532 return findTargetRecurrence(DefOp.getReg(), TargetRegs, RC); in findTargetRecurrence() 1538 return findTargetRecurrence(DefOp.getReg(), TargetRegs, RC); in findTargetRecurrence() 1839 const MachineOperand DefOp = Def->getOperand(DefIdx); in getNextSourceFromBitcast() local 1840 if (DefOp.getSubReg() != DefSubReg) in getNextSourceFromBitcast() 1868 for (const MachineInstr &UseMI : MRI.use_nodbg_instructions(DefOp.getReg())) { in getNextSourceFromBitcast()
|
| H A D | SplitKit.cpp | 449 for (const MachineOperand &DefOp : DefMI->defs()) { in addDeadDef() local 450 Register R = DefOp.getReg(); in addDeadDef() 453 if (unsigned SR = DefOp.getSubReg()) in addDeadDef()
|
| H A D | MachinePipeliner.cpp | 390 MachineOperand &DefOp = PI.getOperand(0); in preprocessPhiNodes() local 391 assert(DefOp.getSubReg() == 0); in preprocessPhiNodes() 392 auto *RC = MRI.getRegClass(DefOp.getReg()); in preprocessPhiNodes()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/ |
| H A D | X86SpeculativeLoadHardening.cpp | 1207 if (const MachineOperand *DefOp = MI.findRegisterDefOperand(X86::EFLAGS)) { in isEFLAGSDefLive() local 1208 return !DefOp->isDead(); in isEFLAGSDefLive() 1218 if (MachineOperand *DefOp = MI.findRegisterDefOperand(X86::EFLAGS)) { in isEFLAGSLive() local 1220 if (DefOp->isDead()) in isEFLAGSLive() 1965 auto &DefOp = MI.getOperand(0); in hardenPostLoad() local 1966 Register OldDefReg = DefOp.getReg(); in hardenPostLoad() 1973 DefOp.setReg(UnhardenedReg); in hardenPostLoad()
|
| H A D | X86DomainReassignment.cpp | 588 for (auto &DefOp : UseMI.defs()) { in buildClosure() local 589 if (!DefOp.isReg()) in buildClosure() 592 Register DefReg = DefOp.getReg(); in buildClosure()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/Hexagon/ |
| H A D | HexagonExpandCondsets.cpp | 226 void predicateAt(const MachineOperand &DefOp, MachineInstr &MI, 855 void HexagonExpandCondsets::predicateAt(const MachineOperand &DefOp, in predicateAt() argument 886 MB.addReg(DefOp.getReg(), getRegState(DefOp), DefOp.getSubReg()); in predicateAt()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/ |
| H A D | MachineTraceMetrics.h | 335 void addLiveIns(const MachineInstr *DefMI, unsigned DefOp,
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
| H A D | SIFoldOperands.cpp | 568 MachineOperand &DefOp = Def->getOperand(1); in tryToFoldACImm() local 569 if (DefOp.isImm() && TII->isInlineConstant(DefOp, OpTy) && in tryToFoldACImm() 570 TII->isOperandLegal(*UseMI, UseOpIdx, &DefOp)) { in tryToFoldACImm() 571 UseMI->getOperand(UseOpIdx).ChangeToImmediate(DefOp.getImm()); in tryToFoldACImm()
|
| H A D | SIInstrInfo.cpp | 519 MachineOperand &DefOp = Def->getOperand(1); in indirectCopyToAGPR() local 520 assert(DefOp.isReg() || DefOp.isImm()); in indirectCopyToAGPR() 522 if (DefOp.isReg()) { in indirectCopyToAGPR() 527 if (I->modifiesRegister(DefOp.getReg(), &RI)) in indirectCopyToAGPR() 533 DefOp.setIsKill(false); in indirectCopyToAGPR() 538 .add(DefOp); in indirectCopyToAGPR()
|