| /netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
| H A D | TargetInstrInfo.cpp | 654 MachineInstr &LoadMI, in foldMemoryOperand() argument 656 assert(LoadMI.canFoldAsLoad() && "LoadMI isn't foldable!"); in foldMemoryOperand() 672 isLoadFromStackSlot(LoadMI, FrameIndex)) { in foldMemoryOperand() 679 NewMI = foldMemoryOperandImpl(MF, MI, Ops, MI, LoadMI, LIS); in foldMemoryOperand() 687 NewMI->setMemRefs(MF, LoadMI.memoperands()); in foldMemoryOperand() 691 for (MachineInstr::mmo_iterator I = LoadMI.memoperands_begin(), in foldMemoryOperand() 692 E = LoadMI.memoperands_end(); in foldMemoryOperand()
|
| H A D | InlineSpiller.cpp | 232 MachineInstr *LoadMI = nullptr); 822 MachineInstr *LoadMI) { in foldMemoryOperand() argument 863 if (LoadMI && MO.isDef()) in foldMemoryOperand() 894 LoadMI ? TII.foldMemoryOperand(*MI, FoldOps, *LoadMI, &LIS) in foldMemoryOperand()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/ |
| H A D | AArch64LoadStoreOptimizer.cpp | 1237 MachineInstr &LoadMI = *I; in findMatchingStore() local 1238 Register BaseReg = getLdStBaseOp(LoadMI).getReg(); in findMatchingStore() 1266 if (MI.mayStore() && isMatchingStore(LoadMI, MI) && in findMatchingStore() 1268 isLdOffsetInRangeOfSt(LoadMI, MI, TII) && in findMatchingStore() 1286 if (MI.mayStore() && LoadMI.mayAlias(AA, MI, /*UseTBAA*/ false)) in findMatchingStore()
|
| H A D | AArch64FastISel.cpp | 4437 const auto *LoadMI = MI; in optimizeIntExtLoad() local 4438 if (LoadMI->getOpcode() == TargetOpcode::COPY && in optimizeIntExtLoad() 4439 LoadMI->getOperand(1).getSubReg() == AArch64::sub_32) { in optimizeIntExtLoad() 4441 LoadMI = MRI.getUniqueVRegDef(LoadReg); in optimizeIntExtLoad() 4442 assert(LoadMI && "Expected valid instruction"); in optimizeIntExtLoad() 4444 if (!(IsZExt && isZExtLoad(LoadMI)) && !(!IsZExt && isSExtLoad(LoadMI))) in optimizeIntExtLoad()
|
| H A D | AArch64InstrInfo.cpp | 4241 MachineInstr &LoadMI = *--InsertPt; in foldMemoryOperandImpl() local 4242 MachineOperand &LoadDst = LoadMI.getOperand(0); in foldMemoryOperandImpl() 4246 return &LoadMI; in foldMemoryOperandImpl()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/SystemZ/ |
| H A D | SystemZInstrInfo.h | 284 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,
|
| H A D | SystemZInstrInfo.cpp | 1328 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl() argument
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/GISel/ |
| H A D | AArch64InstructionSelector.cpp | 2418 auto *LoadMI = emitLoadFromConstantPool(FPImm, MIB); in select() local 2419 if (!LoadMI) { in select() 2423 MIB.buildCopy({DefReg}, {LoadMI->getOperand(0).getReg()}); in select() 2966 auto *LoadMI = getOpcodeDef(TargetOpcode::G_LOAD, SrcReg, MRI); in select() local 2969 if (LoadMI && IsGPR) { in select() 2970 const MachineMemOperand *MemOp = *LoadMI->memoperands_begin(); in select() 4023 MachineInstr *LoadMI = nullptr; in emitLoadFromConstantPool() local 4028 LoadMI = in emitLoadFromConstantPool() 4035 LoadMI = in emitLoadFromConstantPool() 4042 LoadMI = in emitLoadFromConstantPool() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/ |
| H A D | TargetInstrInfo.h | 1093 MachineInstr &LoadMI, 1209 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/PowerPC/ |
| H A D | PPCMIPeephole.cpp | 535 MachineInstr *LoadMI = MRI->getVRegDef(FeedReg1); in simplifyCode() local 536 if (LoadMI && LoadMI->getOpcode() == PPC::LXVDSX) in simplifyCode()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/ |
| H A D | X86InstrInfo.h | 391 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,
|
| H A D | X86InstrInfo.cpp | 5759 static bool isNonFoldablePartialRegisterLoad(const MachineInstr &LoadMI, in isNonFoldablePartialRegisterLoad() argument 5762 unsigned Opc = LoadMI.getOpcode(); in isNonFoldablePartialRegisterLoad() 5766 MF.getRegInfo().getRegClass(LoadMI.getOperand(0).getReg()); in isNonFoldablePartialRegisterLoad() 5990 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl() argument 6001 unsigned NumOps = LoadMI.getDesc().getNumOperands(); in foldMemoryOperandImpl() 6003 if (isLoadFromStackSlot(LoadMI, FrameIndex)) { in foldMemoryOperandImpl() 6004 if (isNonFoldablePartialRegisterLoad(LoadMI, MI, MF)) in foldMemoryOperandImpl() 6020 if (LoadMI.hasOneMemOperand()) in foldMemoryOperandImpl() 6021 Alignment = (*LoadMI.memoperands_begin())->getAlign(); in foldMemoryOperandImpl() 6023 switch (LoadMI.getOpcode()) { in foldMemoryOperandImpl() [all …]
|
| H A D | X86FastISel.cpp | 793 MachineInstrBuilder LoadMI = in handleConstantAddresses() local 795 addFullAddress(LoadMI, StubAM); in handleConstantAddresses()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/GlobalISel/ |
| H A D | CombinerHelper.cpp | 671 if (auto *LoadMI = getOpcodeDef(TargetOpcode::G_SEXTLOAD, LoadUser, MRI)) { in matchSextTruncSextLoad() local 672 const auto &MMO = **LoadMI->memoperands_begin(); in matchSextTruncSextLoad()
|