| /openbsd-src/gnu/llvm/llvm/lib/Transforms/Scalar/ |
| H A D | MergeICmps.cpp | 74 BCEAtom(GetElementPtrInst *GEP, LoadInst *LoadI, int BaseId, APInt Offset) in BCEAtom() 75 : GEP(GEP), LoadI(LoadI), BaseId(BaseId), Offset(Offset) {} in BCEAtom() 85 LoadI = that.LoadI; in operator =() 106 LoadInst *LoadI = nullptr; member 134 auto *const LoadI = dyn_cast<LoadInst>(Val); in visitICmpLoadOperand() local 135 if (!LoadI) in visitICmpLoadOperand() 138 if (LoadI->isUsedOutsideOfBlock(LoadI->getParent())) { in visitICmpLoadOperand() 143 if (!LoadI->isSimple()) { in visitICmpLoadOperand() 147 Value *Addr = LoadI->getOperand(0); in visitICmpLoadOperand() 152 const auto &DL = LoadI->getModule()->getDataLayout(); in visitICmpLoadOperand() [all …]
|
| H A D | JumpThreading.cpp | 1211 if (LoadInst *LoadI = dyn_cast<LoadInst>(SimplifyValue)) in processBlock() local 1212 if (simplifyPartiallyRedundantLoad(LoadI)) in processBlock() 1322 bool JumpThreadingPass::simplifyPartiallyRedundantLoad(LoadInst *LoadI) { in simplifyPartiallyRedundantLoad() argument 1324 if (!LoadI->isUnordered()) return false; in simplifyPartiallyRedundantLoad() 1328 BasicBlock *LoadBB = LoadI->getParent(); in simplifyPartiallyRedundantLoad() 1338 Value *LoadedPtr = LoadI->getOperand(0); in simplifyPartiallyRedundantLoad() 1347 BasicBlock::iterator BBIt(LoadI); in simplifyPartiallyRedundantLoad() 1350 LoadI, LoadBB, BBIt, DefMaxInstsToScan, AA, &IsLoadCSE)) { in simplifyPartiallyRedundantLoad() 1356 combineMetadataForCSE(NLoadI, LoadI, false); in simplifyPartiallyRedundantLoad() 1361 if (AvailableVal == LoadI) in simplifyPartiallyRedundantLoad() [all …]
|
| H A D | DeadStoreElimination.cpp | 1910 if (auto *LoadI = dyn_cast<LoadInst>(Store->getOperand(0))) { in storeIsNoop() local 1911 if (LoadI->getPointerOperand() == Store->getOperand(1)) { in storeIsNoop() 1913 auto *LoadAccess = MSSA.getMemoryAccess(LoadI)->getDefiningAccess(); in storeIsNoop()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/AArch64/ |
| H A D | AArch64LoadStoreOptimizer.cpp | 165 promoteLoadFromStore(MachineBasicBlock::iterator LoadI, 1068 AArch64LoadStoreOpt::promoteLoadFromStore(MachineBasicBlock::iterator LoadI, in promoteLoadFromStore() argument 1071 next_nodbg(LoadI, LoadI->getParent()->end()); in promoteLoadFromStore() 1073 int LoadSize = TII->getMemScale(*LoadI); in promoteLoadFromStore() 1075 Register LdRt = getLdStRegOp(*LoadI).getReg(); in promoteLoadFromStore() 1090 LoadI->getIterator())) { in promoteLoadFromStore() 1097 LLVM_DEBUG(LoadI->print(dbgs())); in promoteLoadFromStore() 1099 LoadI->eraseFromParent(); in promoteLoadFromStore() 1104 BuildMI(*LoadI->getParent(), LoadI, LoadI->getDebugLoc(), in promoteLoadFromStore() 1109 .setMIFlags(LoadI->getFlags()); in promoteLoadFromStore() [all …]
|
| H A D | AArch64FalkorHWPFFix.cpp | 154 LoadInst *LoadI = dyn_cast<LoadInst>(&I); in runOnLoop() local 155 if (!LoadI) in runOnLoop() 158 Value *PtrValue = LoadI->getPointerOperand(); in runOnLoop() 167 LoadI->setMetadata(FALKOR_STRIDED_ACCESS_MD, in runOnLoop() 168 MDNode::get(LoadI->getContext(), {})); in runOnLoop()
|
| /openbsd-src/gnu/llvm/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombinePHI.cpp | 120 if (LoadInst *LoadI = dyn_cast<LoadInst>(U)) { in foldIntegerTypedPHI() local 121 Ptr = LoadI->getPointerOperand(); in foldIntegerTypedPHI() 176 auto *LoadI = dyn_cast<LoadInst>(Arg); in foldIntegerTypedPHI() local 177 if (!LoadI) in foldIntegerTypedPHI() 180 if (!LoadI->hasOneUse()) in foldIntegerTypedPHI() 186 AvailablePtrVals.emplace_back(LoadI); in foldIntegerTypedPHI() 263 LoadInst *LoadI = dyn_cast<LoadInst>(IncomingVal); in foldIntegerTypedPHI() local 266 (LoadI && LoadI->hasOneUse())) && in foldIntegerTypedPHI()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/Hexagon/ |
| H A D | HexagonVLIWPacketizer.cpp | 1530 bool LoadI = I.mayLoad(), StoreI = I.mayStore(); in isLegalToPacketizeTogether() local 1538 if (LoadJ && LoadI && HII->isPureSlot0(J)) { in isLegalToPacketizeTogether() 1545 (StoreJ && LoadI && !NVStoreJ)) && in isLegalToPacketizeTogether() 1553 if (StoreJ && LoadI && alias(J, I)) { in isLegalToPacketizeTogether() 1559 if (!LoadJ || (!LoadI && !StoreI)) { in isLegalToPacketizeTogether()
|
| /openbsd-src/gnu/llvm/llvm/lib/CodeGen/ |
| H A D | SelectOptimize.cpp | 735 static bool isSafeToSinkLoad(Instruction *LoadI, Instruction *SI) { in isSafeToSinkLoad() argument 737 if (LoadI->getParent() != SI->getParent()) in isSafeToSinkLoad() 739 auto It = LoadI->getIterator(); in isSafeToSinkLoad()
|
| /openbsd-src/gnu/llvm/llvm/lib/Transforms/IPO/ |
| H A D | AttributorAttributes.cpp | 1656 if (auto *LoadI = dyn_cast<LoadInst>(Usr)) { in updateImpl() local 1664 if (!handleAccess(A, *LoadI, /* Content */ nullptr, AK, in updateImpl() 1666 *LoadI->getType())) in updateImpl() 1686 BasicBlock *BB = LoadI->getParent(); in updateImpl() 1692 if (IsImpactedInRange(LoadI->getNextNonDebugInstruction(), &IntrI)) in updateImpl() 1707 if (IsImpactedInRange(LoadI->getNextNonDebugInstruction(), in updateImpl() 1717 for (const Use &LoadU : LoadI->uses()) { in updateImpl() 1740 << *Assumption.second << ": " << *LoadI in updateImpl() 1745 OffsetInfoMap[CurPtr].Offsets, Changed, *LoadI->getType()); in updateImpl()
|
| H A D | OpenMPOpt.cpp | 1051 LoadInst *LoadI = new LoadInst( in mergeParallelRegions() local 1053 UsrI->replaceUsesOfWith(&I, LoadI); in mergeParallelRegions() 3766 LoadInst *LoadI = new LoadInst(I.getType(), SharedMem, in insertInstructionGuardsHelper() local 3772 UsrI->replaceUsesOfWith(&I, LoadI); in insertInstructionGuardsHelper()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/SystemZ/ |
| H A D | SystemZISelLowering.cpp | 931 if (auto *LoadI = dyn_cast<LoadInst>(StoreI->getValueOperand())) in supportedAddressingMode() local 932 if (LoadI->hasOneUse() && LoadI->getParent() == I->getParent()) in supportedAddressingMode() 934 return getLoadStoreAddrMode(HasVector, LoadI->getType()); in supportedAddressingMode()
|