/freebsd-src/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUPostLegalizerCombiner.cpp | 405 MachineInstr *LoadMI = MRI.getVRegDef(LoadReg); matchCombineSignExtendInReg() local
|
/freebsd-src/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | InlineSpiller.cpp | 890 MachineInstr *LoadMI) { in foldMemoryOperand() argument
|
H A D | TargetInstrInfo.cpp | 731 MachineInstr &LoadMI, in foldMemoryOperand() argument
|
/freebsd-src/contrib/llvm-project/llvm/lib/Target/PowerPC/ |
H A D | PPCMIPeephole.cpp | 651 MachineInstr *LoadMI = MRI->getVRegDef(FeedReg1); simplifyCode() local
|
/freebsd-src/contrib/llvm-project/llvm/lib/CodeGen/GlobalISel/ |
H A D | CombinerHelper.cpp | 437 ChoosePreferredUse(MachineInstr & LoadMI,PreferredTuple & CurrentUse,const LLT TyForCandidate,unsigned OpcodeForCandidate,MachineInstr * MIForCandidate) ChoosePreferredUse() argument 556 GAnyLoad *LoadMI = dyn_cast<GAnyLoad>(&MI); matchCombineExtendingLoads() local 759 GAnyLoad *LoadMI = dyn_cast<GAnyLoad>(MRI.getVRegDef(SrcReg)); matchCombineLoadWithAndMask() local 854 if (auto *LoadMI = getOpcodeDef<GSExtLoad>(LoadUser, MRI)) { matchSextTruncSextLoad() local 1173 auto *LoadMI = getOpcodeDef<GLoad>(MI.getOperand(1).getReg(), MRI); matchCombineExtractedVectorLoad() local [all...] |
H A D | LegalizerHelper.cpp | 1314 auto &LoadMI = cast<GLoad>(MI); narrowScalar() local 1332 auto &LoadMI = cast<GExtLoad>(MI); narrowScalar() local 3297 lowerLoad(GAnyLoad & LoadMI) lowerLoad() argument [all...] |
/freebsd-src/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64LoadStoreOptimizer.cpp | 1296 MachineInstr &LoadMI = *I; findMatchingStore() local
|
H A D | AArch64FastISel.cpp | 4517 const auto *LoadMI = MI; optimizeIntExtLoad() local
|
H A D | AArch64InstrInfo.cpp | 5602 MachineInstr &LoadMI = *--InsertPt; foldMemoryOperandImpl() local
|
/freebsd-src/contrib/llvm-project/llvm/lib/Target/SystemZ/ |
H A D | SystemZInstrInfo.cpp | 1350 foldMemoryOperandImpl(MachineFunction & MF,MachineInstr & MI,ArrayRef<unsigned> Ops,MachineBasicBlock::iterator InsertPt,MachineInstr & LoadMI,LiveIntervals * LIS) const foldMemoryOperandImpl() argument
|
/freebsd-src/contrib/llvm-project/llvm/lib/Target/AArch64/GISel/ |
H A D | AArch64InstructionSelector.cpp | 2682 auto *LoadMI = emitLoadFromConstantPool(FPImm, MIB); select() local 3330 auto *LoadMI = getOpcodeDef(TargetOpcode::G_LOAD, SrcReg, MRI); select() local 4335 MachineInstr *LoadMI = nullptr; emitLoadFromConstantPool() local [all...] |
/freebsd-src/contrib/llvm-project/llvm/lib/Target/X86/ |
H A D | X86InstrInfo.cpp | 7551 isNonFoldablePartialRegisterLoad(const MachineInstr & LoadMI,const MachineInstr & UserMI,const MachineFunction & MF) isNonFoldablePartialRegisterLoad() argument 7975 foldMemoryOperandImpl(MachineFunction & MF,MachineInstr & MI,ArrayRef<unsigned> Ops,MachineBasicBlock::iterator InsertPt,MachineInstr & LoadMI,LiveIntervals * LIS) const foldMemoryOperandImpl() argument [all...] |
H A D | X86FastISel.cpp | 791 MachineInstrBuilder LoadMI = in handleConstantAddresses() local
|