Home
last modified time | relevance | path

Searched refs:LoadMI (Results 1 – 14 of 14) sorted by relevance

/netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/
H A DTargetInstrInfo.cpp654 MachineInstr &LoadMI, in foldMemoryOperand() argument
656 assert(LoadMI.canFoldAsLoad() && "LoadMI isn't foldable!"); in foldMemoryOperand()
672 isLoadFromStackSlot(LoadMI, FrameIndex)) { in foldMemoryOperand()
679 NewMI = foldMemoryOperandImpl(MF, MI, Ops, MI, LoadMI, LIS); in foldMemoryOperand()
687 NewMI->setMemRefs(MF, LoadMI.memoperands()); in foldMemoryOperand()
691 for (MachineInstr::mmo_iterator I = LoadMI.memoperands_begin(), in foldMemoryOperand()
692 E = LoadMI.memoperands_end(); in foldMemoryOperand()
H A DInlineSpiller.cpp232 MachineInstr *LoadMI = nullptr);
822 MachineInstr *LoadMI) { in foldMemoryOperand() argument
863 if (LoadMI && MO.isDef()) in foldMemoryOperand()
894 LoadMI ? TII.foldMemoryOperand(*MI, FoldOps, *LoadMI, &LIS) in foldMemoryOperand()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/
H A DAArch64LoadStoreOptimizer.cpp1237 MachineInstr &LoadMI = *I; in findMatchingStore() local
1238 Register BaseReg = getLdStBaseOp(LoadMI).getReg(); in findMatchingStore()
1266 if (MI.mayStore() && isMatchingStore(LoadMI, MI) && in findMatchingStore()
1268 isLdOffsetInRangeOfSt(LoadMI, MI, TII) && in findMatchingStore()
1286 if (MI.mayStore() && LoadMI.mayAlias(AA, MI, /*UseTBAA*/ false)) in findMatchingStore()
H A DAArch64FastISel.cpp4437 const auto *LoadMI = MI; in optimizeIntExtLoad() local
4438 if (LoadMI->getOpcode() == TargetOpcode::COPY && in optimizeIntExtLoad()
4439 LoadMI->getOperand(1).getSubReg() == AArch64::sub_32) { in optimizeIntExtLoad()
4441 LoadMI = MRI.getUniqueVRegDef(LoadReg); in optimizeIntExtLoad()
4442 assert(LoadMI && "Expected valid instruction"); in optimizeIntExtLoad()
4444 if (!(IsZExt && isZExtLoad(LoadMI)) && !(!IsZExt && isSExtLoad(LoadMI))) in optimizeIntExtLoad()
H A DAArch64InstrInfo.cpp4241 MachineInstr &LoadMI = *--InsertPt; in foldMemoryOperandImpl() local
4242 MachineOperand &LoadDst = LoadMI.getOperand(0); in foldMemoryOperandImpl()
4246 return &LoadMI; in foldMemoryOperandImpl()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/SystemZ/
H A DSystemZInstrInfo.h284 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,
H A DSystemZInstrInfo.cpp1328 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl() argument
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/GISel/
H A DAArch64InstructionSelector.cpp2418 auto *LoadMI = emitLoadFromConstantPool(FPImm, MIB); in select() local
2419 if (!LoadMI) { in select()
2423 MIB.buildCopy({DefReg}, {LoadMI->getOperand(0).getReg()}); in select()
2966 auto *LoadMI = getOpcodeDef(TargetOpcode::G_LOAD, SrcReg, MRI); in select() local
2969 if (LoadMI && IsGPR) { in select()
2970 const MachineMemOperand *MemOp = *LoadMI->memoperands_begin(); in select()
4023 MachineInstr *LoadMI = nullptr; in emitLoadFromConstantPool() local
4028 LoadMI = in emitLoadFromConstantPool()
4035 LoadMI = in emitLoadFromConstantPool()
4042 LoadMI = in emitLoadFromConstantPool()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/
H A DTargetInstrInfo.h1093 MachineInstr &LoadMI,
1209 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/PowerPC/
H A DPPCMIPeephole.cpp535 MachineInstr *LoadMI = MRI->getVRegDef(FeedReg1); in simplifyCode() local
536 if (LoadMI && LoadMI->getOpcode() == PPC::LXVDSX) in simplifyCode()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/
H A DX86InstrInfo.h391 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI,
H A DX86InstrInfo.cpp5759 static bool isNonFoldablePartialRegisterLoad(const MachineInstr &LoadMI, in isNonFoldablePartialRegisterLoad() argument
5762 unsigned Opc = LoadMI.getOpcode(); in isNonFoldablePartialRegisterLoad()
5766 MF.getRegInfo().getRegClass(LoadMI.getOperand(0).getReg()); in isNonFoldablePartialRegisterLoad()
5990 MachineBasicBlock::iterator InsertPt, MachineInstr &LoadMI, in foldMemoryOperandImpl() argument
6001 unsigned NumOps = LoadMI.getDesc().getNumOperands(); in foldMemoryOperandImpl()
6003 if (isLoadFromStackSlot(LoadMI, FrameIndex)) { in foldMemoryOperandImpl()
6004 if (isNonFoldablePartialRegisterLoad(LoadMI, MI, MF)) in foldMemoryOperandImpl()
6020 if (LoadMI.hasOneMemOperand()) in foldMemoryOperandImpl()
6021 Alignment = (*LoadMI.memoperands_begin())->getAlign(); in foldMemoryOperandImpl()
6023 switch (LoadMI.getOpcode()) { in foldMemoryOperandImpl()
[all …]
H A DX86FastISel.cpp793 MachineInstrBuilder LoadMI = in handleConstantAddresses() local
795 addFullAddress(LoadMI, StubAM); in handleConstantAddresses()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/GlobalISel/
H A DCombinerHelper.cpp671 if (auto *LoadMI = getOpcodeDef(TargetOpcode::G_SEXTLOAD, LoadUser, MRI)) { in matchSextTruncSextLoad() local
672 const auto &MMO = **LoadMI->memoperands_begin(); in matchSextTruncSextLoad()