| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/ |
| H A D | X86AvoidStoreForwardingBlocks.cpp | 102 void breakBlockedCopies(MachineInstr *LoadInst, MachineInstr *StoreInst, 105 void buildCopies(int Size, MachineInstr *LoadInst, int64_t LdDispImm, 109 void buildCopy(MachineInstr *LoadInst, unsigned NLoadOpcode, int64_t LoadDisp, 339 findPotentialBlockers(MachineInstr *LoadInst) { in findPotentialBlockers() argument 343 for (auto PBInst = std::next(MachineBasicBlock::reverse_iterator(LoadInst)), in findPotentialBlockers() 344 E = LoadInst->getParent()->rend(); in findPotentialBlockers() 361 MachineBasicBlock *MBB = LoadInst->getParent(); in findPotentialBlockers() 385 void X86AvoidSFBPass::buildCopy(MachineInstr *LoadInst, unsigned NLoadOpcode, in buildCopy() argument 390 MachineOperand &LoadBase = getBaseOperand(LoadInst); in buildCopy() 392 MachineBasicBlock *MBB = LoadInst->getParent(); in buildCopy() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/ARM/ |
| H A D | ARMParallelDSP.cpp | 57 using MemInstList = SmallVectorImpl<LoadInst*>; 69 SmallVector<LoadInst*, 2> VecLd; // Container for loads to widen. 75 return isa<LoadInst>(LHS) && isa<LoadInst>(RHS); in HasTwoLoadInputs() 78 LoadInst *getBaseLoad() const { in getBaseLoad() 199 LoadInst *NewLd = nullptr; 200 SmallVector<LoadInst*, 4> Loads; 203 WidenedLoad(SmallVectorImpl<LoadInst*> &Lds, LoadInst *Wide) in WidenedLoad() 207 LoadInst *getLoad() { in getLoad() 219 std::map<LoadInst*, LoadInst*> LoadPairs; 220 SmallPtrSet<LoadInst*, 4> OffsetLoads; [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/IPO/ |
| H A D | ArgumentPromotion.cpp | 129 std::map<std::pair<Argument *, IndicesVector>, LoadInst *> OriginalLoads; in doPromotion() 166 if (LoadInst *L = dyn_cast<LoadInst>(UI)) in doPromotion() 188 LoadInst *OrigLoad; in doPromotion() 189 if (LoadInst *L = dyn_cast<LoadInst>(UI)) in doPromotion() 193 OrigLoad = cast<LoadInst>(UI->user_back()); in doPromotion() 289 LoadInst *OrigLoad = in doPromotion() 313 LoadInst *newLoad = in doPromotion() 430 if (LoadInst *LI = dyn_cast<LoadInst>(I->user_back())) { in doPromotion() 466 LoadInst *L = cast<LoadInst>(GEP->user_back()); in doPromotion() 624 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in isSafeToPromoteArgument() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Utils/ |
| H A D | Local.h | 53 class LoadInst; variable 251 LoadInst *LI, DIBuilder &Builder); 377 void copyMetadataForLoad(LoadInst &Dest, const LoadInst &Source); 411 void copyNonnullMetadata(const LoadInst &OldLI, MDNode *N, LoadInst &NewLI); 417 void copyRangeMetadata(const DataLayout &DL, const LoadInst &OldLI, MDNode *N, 418 LoadInst &NewLI);
|
| H A D | VNCoercion.h | 27 class LoadInst; variable 62 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI, 87 Value *getLoadValueForLoad(LoadInst *SrcVal, unsigned Offset, Type *LoadTy,
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPUAnnotateUniformValues.cpp | 51 void visitLoadInst(LoadInst &I); 52 bool isClobberedInFunction(LoadInst * Load); 73 bool AMDGPUAnnotateUniformValues::isClobberedInFunction(LoadInst * Load) { in isClobberedInFunction() 83 void AMDGPUAnnotateUniformValues::visitLoadInst(LoadInst &I) { in visitLoadInst() 87 auto isGlobalLoad = [&](LoadInst &Load)->bool { in visitLoadInst()
|
| H A D | AMDGPULateCodeGenPrepare.cpp | 77 bool canWidenScalarExtLoad(LoadInst &LI) const; 78 bool visitLoadInst(LoadInst &LI); 106 bool AMDGPULateCodeGenPrepare::canWidenScalarExtLoad(LoadInst &LI) const { in canWidenScalarExtLoad() 130 bool AMDGPULateCodeGenPrepare::visitLoadInst(LoadInst &LI) { in visitLoadInst() 171 LoadInst *NewLd = IRB.CreateAlignedLoad(IRB.getInt32Ty(), NewPtr, Align(4)); in visitLoadInst()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/NVPTX/ |
| H A D | NVPTXLowerArgs.cpp | 174 if (auto *LI = dyn_cast<LoadInst>(I.OldInstruction)) { in convertToParamAS() 240 if (isa<GetElementPtrInst>(V) || isa<BitCastInst>(V) || isa<LoadInst>(V)) in handleByValParam() 258 if (!isa<LoadInst>(V)) in handleByValParam() 295 LoadInst *LI = in handleByValParam() 296 new LoadInst(StructType, ArgInParam, Arg->getName(), in handleByValParam() 336 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in runOnKernelFunction()
|
| H A D | NVPTXLowerAggrCopies.cpp | 60 SmallVector<LoadInst *, 4> AggrLoads; in runOnFunction() 72 if (LoadInst *LI = dyn_cast<LoadInst>(II)) { in runOnFunction() 105 for (LoadInst *LI : AggrLoads) { in runOnFunction()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Scalar/ |
| H A D | GVN.h | 48 class LoadInst; variable 311 bool processLoad(LoadInst *L); 312 bool processNonLocalLoad(LoadInst *L); 318 bool AnalyzeLoadAvailability(LoadInst *Load, MemDepResult DepInfo, 324 void AnalyzeLoadAvailability(LoadInst *Load, LoadDepVect &Deps, 328 bool PerformLoadPRE(LoadInst *Load, AvailValInBlkVect &ValuesPerBlock, 334 bool performLoopLoadPRE(LoadInst *Load, AvailValInBlkVect &ValuesPerBlock, 340 LoadInst *Load, AvailValInBlkVect &ValuesPerBlock,
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/ |
| H A D | SSAUpdater.cpp | 327 if (const LoadInst *LI = dyn_cast<LoadInst>(Insts[0])) in LoadAndStorePromoter() 349 SmallVector<LoadInst *, 32> LiveInLoads; in run() 368 LiveInLoads.push_back(cast<LoadInst>(User)); in run() 387 LiveInLoads.push_back(cast<LoadInst>(I)); in run() 399 if (LoadInst *L = dyn_cast<LoadInst>(&I)) { in run() 433 for (LoadInst *ALoad : LiveInLoads) { in run() 466 replaceLoadWithValue(cast<LoadInst>(User), NewVal); in run()
|
| H A D | PromoteMemoryToRegister.cpp | 67 if (const LoadInst *LI = dyn_cast<LoadInst>(U)) { in isAllocaPromotable() 140 LoadInst *LI = cast<LoadInst>(User); in AnalyzeAlloca() 189 return (isa<LoadInst>(I) && isa<AllocaInst>(I->getOperand(0))) || in isInterestingInstruction() 301 static void addAssumeNonNull(AssumptionCache *AC, LoadInst *LI) { in addAssumeNonNull() 318 if (isa<LoadInst>(I) || isa<StoreInst>(I)) in removeIntrinsicUsers() 369 LoadInst *LI = cast<LoadInst>(UserInst); in rewriteSingleStoreAlloca() 480 LoadInst *LI = dyn_cast<LoadInst>(U); in promoteSingleBlockAlloca() 818 if (LoadInst *LI = dyn_cast<LoadInst>(I)) in ComputeLiveInBlocks() 947 if (LoadInst *LI = dyn_cast<LoadInst>(I)) { in RenamePass()
|
| H A D | GlobalStatus.cpp | 87 if (const LoadInst *LI = dyn_cast<LoadInst>(I)) { in analyzeGlobalAux() 122 } else if (isa<LoadInst>(StoredVal) && in analyzeGlobalAux() 123 cast<LoadInst>(StoredVal)->getOperand(0) == GV) { in analyzeGlobalAux()
|
| H A D | RelLookupTableConverter.cpp | 44 LoadInst *Load = dyn_cast<LoadInst>(GEP->use_begin()->getUser()); in shouldConvertToRelLookupTable() 130 LoadInst *Load = cast<LoadInst>(GEP->use_begin()->getUser()); in convertToRelLookupTable()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Analysis/ |
| H A D | MemDerefPrinter.cpp | 58 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in runOnFunction() 93 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in run()
|
| H A D | MemoryDependenceAnalysis.cpp | 118 if (const LoadInst *LI = dyn_cast<LoadInst>(Inst)) { in GetLocation() 252 if (auto *LI = dyn_cast<LoadInst>(QueryInst)) { in getPointerDependencyFrom() 283 MemoryDependenceResults::getInvariantGroupPointerDependency(LoadInst *LI, in getInvariantGroupPointerDependency() 346 if ((isa<LoadInst>(U) || in getInvariantGroupPointerDependency() 412 LoadInst *LI = dyn_cast<LoadInst>(QueryInst); in getSimplePointerDependencyFrom() 420 if (auto *LI = dyn_cast<LoadInst>(I)) in getSimplePointerDependencyFrom() 430 return !isa<LoadInst>(I) && !isa<StoreInst>(I) && I->mayReadOrWriteMemory(); in getSimplePointerDependencyFrom() 484 if (LoadInst *LI = dyn_cast<LoadInst>(Inst)) { in getSimplePointerDependencyFrom() 845 bool isLoad = isa<LoadInst>(QueryInst); in getNonLocalPointerDependency() 872 if (LoadInst *LI = dyn_cast<LoadInst>(Inst)) { in getNonLocalPointerDependency() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Analysis/ |
| H A D | Loads.h | 25 class LoadInst; variable 83 bool isDereferenceableAndAlignedInLoop(LoadInst *LI, Loop *L, 129 Value *FindAvailableLoadedValue(LoadInst *Load, 141 Value *FindAvailableLoadedValue(LoadInst *Load, AAResults &AA, bool *IsLoadCSE,
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/ |
| H A D | LowerAtomic.cpp | 30 LoadInst *Orig = Builder.CreateLoad(Val->getType(), Ptr); in LowerAtomicCmpXchgInst() 48 LoadInst *Orig = Builder.CreateLoad(Val->getType(), Ptr); in lowerAtomicRMWInst() 108 static bool LowerLoadInst(LoadInst *LI) { in LowerLoadInst() 127 else if (LoadInst *LI = dyn_cast<LoadInst>(&Inst)) { in runOnBasicBlock()
|
| H A D | LoopLoadElimination.cpp | 90 LoadInst *Load; 93 StoreToLoadForwardingCandidate(LoadInst *Load, StoreInst *Store) in StoreToLoadForwardingCandidate() 157 static bool isLoadConditional(LoadInst *Load, Loop *L) { in isLoadConditional() 195 if (isa<LoadInst>(Source)) in findStoreToLoadDependences() 197 if (isa<LoadInst>(Destination)) in findStoreToLoadDependences() 213 auto *Load = dyn_cast<LoadInst>(Destination); in findStoreToLoadDependences() 263 DenseMap<LoadInst *, const StoreToLoadForwardingCandidate *>; in removeDependencesFromMultipleStores() 344 LoadInst *LastLoad = in findPointersWrittenOnForwardingPath() 436 Value *Initial = new LoadInst( in propagateStoredValueToLoadUsers()
|
| H A D | SROA.cpp | 776 void visitLoadInst(LoadInst &LI) { in visitLoadInst() 975 if (LoadInst *LI = dyn_cast<LoadInst>(I)) { in hasUnsafePHIOrSelectUse() 1148 if (LoadInst *LI = dyn_cast<LoadInst>(U->getUser())) { in findCommonType() 1211 LoadInst *LI = dyn_cast<LoadInst>(U); in isSafePHIToSpeculate() 1269 LoadInst *SomeLoad = cast<LoadInst>(PN.user_back()); in speculatePHINodeLoads() 1283 LoadInst *LI = cast<LoadInst>(PN.user_back()); in speculatePHINodeLoads() 1306 LoadInst *Load = PredBuilder.CreateAlignedLoad( in speculatePHINodeLoads() 1339 LoadInst *LI = dyn_cast<LoadInst>(U); in isSafeSelectToSpeculate() 1365 LoadInst *LI = cast<LoadInst>(SI.user_back()); in speculateSelectInstLoads() 1369 LoadInst *TL = IRB.CreateLoad(LI->getType(), TV, in speculateSelectInstLoads() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombineLoadStoreAlloca.cpp | 60 if (auto *LI = dyn_cast<LoadInst>(I)) { in isOnlyCopiedFromConstantMemory() 257 if (LoadInst *Load = dyn_cast<LoadInst>(Inst)) { in collectUsers() 282 if (auto *LT = dyn_cast<LoadInst>(I)) { in replace() 285 auto *NewI = new LoadInst(LT->getType(), V, "", LT->isVolatile(), in replace() 459 LoadInst *InstCombinerImpl::combineLoadToNewType(LoadInst &LI, Type *NewTy, in combineLoadToNewType() 472 LoadInst *NewLoad = Builder.CreateAlignedLoad( in combineLoadToNewType() 577 LoadInst &LI) { in combineLoadToOperationType() 609 LoadInst *NewLoad = IC.combineLoadToNewType(LI, CI->getDestTy()); in combineLoadToOperationType() 621 static Instruction *unpackLoadToAggregate(InstCombinerImpl &IC, LoadInst &LI) { in unpackLoadToAggregate() 638 LoadInst *NewLoad = IC.combineLoadToNewType(LI, ST->getTypeAtIndex(0U), in unpackLoadToAggregate() [all …]
|
| H A D | InstCombinePHI.cpp | 119 if (LoadInst *LoadI = dyn_cast<LoadInst>(U)) { in foldIntegerTypedPHI() 174 auto *LoadI = dyn_cast<LoadInst>(Arg); in foldIntegerTypedPHI() 263 LoadInst *LoadI = dyn_cast<LoadInst>(IncomingVal); in foldIntegerTypedPHI() 591 static bool isSafeAndProfitableToSinkLoad(LoadInst *L) { in isSafeAndProfitableToSinkLoad() 609 if (isa<LoadInst>(U)) continue; in isSafeAndProfitableToSinkLoad() 636 LoadInst *FirstLI = cast<LoadInst>(PN.getIncomingValue(0)); in foldPHIArgLoadIntoPHI() 667 LoadInst *LI = dyn_cast<LoadInst>(PN.getIncomingValue(i)); in foldPHIArgLoadIntoPHI() 697 LoadInst *NewLI = in foldPHIArgLoadIntoPHI() 698 new LoadInst(FirstLI->getType(), NewPN, "", isVolatile, LoadAlignment); in foldPHIArgLoadIntoPHI() 718 LoadInst *LI = cast<LoadInst>(PN.getIncomingValue(i)); in foldPHIArgLoadIntoPHI() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
| H A D | InterleavedLoadCombinePass.cpp | 94 LoadInst *findFirstLoad(const std::set<LoadInst *> &LIs); 651 LoadInst *LI; 653 ElementInfo(Polynomial Offset = Polynomial(), LoadInst *LI = nullptr) in ElementInfo() 664 std::set<LoadInst *> LIs; 716 LoadInst *LI = dyn_cast<LoadInst>(V); in compute() 868 static bool computeFromLI(LoadInst *LI, VectorInfo &Result, in computeFromLI() 1101 LoadInst * 1102 InterleavedLoadCombineImpl::findFirstLoad(const std::set<LoadInst *> &LIs) { in findFirstLoad() 1111 return cast<LoadInst>(FLI); in findFirstLoad() 1121 LoadInst *InsertionPoint = InterleavedLoad.front().EI[0].LI; in combine() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Instrumentation/ |
| H A D | ThreadSanitizer.cpp | 422 } else if (LoadInst *L = dyn_cast<LoadInst>(Addr)) { in addrPointsToConstantData() 452 : cast<LoadInst>(I)->getPointerOperand(); in chooseInstructionsToInstrument() 464 ClDistinguishVolatile && (cast<LoadInst>(I)->isVolatile() || in chooseInstructionsToInstrument() 503 if (LoadInst *LI = dyn_cast<LoadInst>(I)) in isAtomic() 551 else if (isa<LoadInst>(Inst) || isa<StoreInst>(Inst)) in sanitizeFunction() 616 : cast<LoadInst>(II.Inst)->getPointerOperand(); in instrumentLoadOrStore() 653 : cast<LoadInst>(II.Inst)->getAlignment(); in instrumentLoadOrStore() 658 : cast<LoadInst>(II.Inst)->isVolatile()); in instrumentLoadOrStore() 743 if (LoadInst *LI = dyn_cast<LoadInst>(I)) { in instrumentAtomic()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Vectorize/ |
| H A D | LoadStoreVectorizer.cpp | 296 if (LoadInst *L = dyn_cast<LoadInst>(I)) in getPointerAddressSpace() 652 bool IsLoadChain = isa<LoadInst>(Chain[0]); in getVectorizablePrefix() 656 assert(isa<LoadInst>(I) && in getVectorizablePrefix() 665 if (isa<LoadInst>(I) || isa<StoreInst>(I)) { in getVectorizablePrefix() 711 auto *MemLoad = dyn_cast<LoadInst>(MemInstr); in getVectorizablePrefix() 712 auto *ChainLoad = dyn_cast<LoadInst>(ChainInstr); in getVectorizablePrefix() 718 auto IsInvariantLoad = [](const LoadInst *LI) -> bool { in getVectorizablePrefix() 800 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in collectInstructions() 978 if (isa<LoadInst>(*Operands.begin())) in vectorizeInstructions() 1145 LoadInst *L0 = cast<LoadInst>(Chain[0]); in vectorizeLoadChain() [all …]
|