Home
last modified time | relevance | path

Searched refs:LoadInst (Results 1 – 25 of 261) sorted by relevance

1234567891011

/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/
H A DX86AvoidStoreForwardingBlocks.cpp102 void breakBlockedCopies(MachineInstr *LoadInst, MachineInstr *StoreInst,
105 void buildCopies(int Size, MachineInstr *LoadInst, int64_t LdDispImm,
109 void buildCopy(MachineInstr *LoadInst, unsigned NLoadOpcode, int64_t LoadDisp,
339 findPotentialBlockers(MachineInstr *LoadInst) { in findPotentialBlockers() argument
343 for (auto PBInst = std::next(MachineBasicBlock::reverse_iterator(LoadInst)), in findPotentialBlockers()
344 E = LoadInst->getParent()->rend(); in findPotentialBlockers()
361 MachineBasicBlock *MBB = LoadInst->getParent(); in findPotentialBlockers()
385 void X86AvoidSFBPass::buildCopy(MachineInstr *LoadInst, unsigned NLoadOpcode, in buildCopy() argument
390 MachineOperand &LoadBase = getBaseOperand(LoadInst); in buildCopy()
392 MachineBasicBlock *MBB = LoadInst->getParent(); in buildCopy()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/ARM/
H A DARMParallelDSP.cpp57 using MemInstList = SmallVectorImpl<LoadInst*>;
69 SmallVector<LoadInst*, 2> VecLd; // Container for loads to widen.
75 return isa<LoadInst>(LHS) && isa<LoadInst>(RHS); in HasTwoLoadInputs()
78 LoadInst *getBaseLoad() const { in getBaseLoad()
199 LoadInst *NewLd = nullptr;
200 SmallVector<LoadInst*, 4> Loads;
203 WidenedLoad(SmallVectorImpl<LoadInst*> &Lds, LoadInst *Wide) in WidenedLoad()
207 LoadInst *getLoad() { in getLoad()
219 std::map<LoadInst*, LoadInst*> LoadPairs;
220 SmallPtrSet<LoadInst*, 4> OffsetLoads;
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/IPO/
H A DArgumentPromotion.cpp129 std::map<std::pair<Argument *, IndicesVector>, LoadInst *> OriginalLoads; in doPromotion()
166 if (LoadInst *L = dyn_cast<LoadInst>(UI)) in doPromotion()
188 LoadInst *OrigLoad; in doPromotion()
189 if (LoadInst *L = dyn_cast<LoadInst>(UI)) in doPromotion()
193 OrigLoad = cast<LoadInst>(UI->user_back()); in doPromotion()
289 LoadInst *OrigLoad = in doPromotion()
313 LoadInst *newLoad = in doPromotion()
430 if (LoadInst *LI = dyn_cast<LoadInst>(I->user_back())) { in doPromotion()
466 LoadInst *L = cast<LoadInst>(GEP->user_back()); in doPromotion()
624 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in isSafeToPromoteArgument()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Utils/
H A DLocal.h53 class LoadInst; variable
251 LoadInst *LI, DIBuilder &Builder);
377 void copyMetadataForLoad(LoadInst &Dest, const LoadInst &Source);
411 void copyNonnullMetadata(const LoadInst &OldLI, MDNode *N, LoadInst &NewLI);
417 void copyRangeMetadata(const DataLayout &DL, const LoadInst &OldLI, MDNode *N,
418 LoadInst &NewLI);
H A DVNCoercion.h27 class LoadInst; variable
62 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI,
87 Value *getLoadValueForLoad(LoadInst *SrcVal, unsigned Offset, Type *LoadTy,
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/
H A DAMDGPUAnnotateUniformValues.cpp51 void visitLoadInst(LoadInst &I);
52 bool isClobberedInFunction(LoadInst * Load);
73 bool AMDGPUAnnotateUniformValues::isClobberedInFunction(LoadInst * Load) { in isClobberedInFunction()
83 void AMDGPUAnnotateUniformValues::visitLoadInst(LoadInst &I) { in visitLoadInst()
87 auto isGlobalLoad = [&](LoadInst &Load)->bool { in visitLoadInst()
H A DAMDGPULateCodeGenPrepare.cpp77 bool canWidenScalarExtLoad(LoadInst &LI) const;
78 bool visitLoadInst(LoadInst &LI);
106 bool AMDGPULateCodeGenPrepare::canWidenScalarExtLoad(LoadInst &LI) const { in canWidenScalarExtLoad()
130 bool AMDGPULateCodeGenPrepare::visitLoadInst(LoadInst &LI) { in visitLoadInst()
171 LoadInst *NewLd = IRB.CreateAlignedLoad(IRB.getInt32Ty(), NewPtr, Align(4)); in visitLoadInst()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/NVPTX/
H A DNVPTXLowerArgs.cpp174 if (auto *LI = dyn_cast<LoadInst>(I.OldInstruction)) { in convertToParamAS()
240 if (isa<GetElementPtrInst>(V) || isa<BitCastInst>(V) || isa<LoadInst>(V)) in handleByValParam()
258 if (!isa<LoadInst>(V)) in handleByValParam()
295 LoadInst *LI = in handleByValParam()
296 new LoadInst(StructType, ArgInParam, Arg->getName(), in handleByValParam()
336 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in runOnKernelFunction()
H A DNVPTXLowerAggrCopies.cpp60 SmallVector<LoadInst *, 4> AggrLoads; in runOnFunction()
72 if (LoadInst *LI = dyn_cast<LoadInst>(II)) { in runOnFunction()
105 for (LoadInst *LI : AggrLoads) { in runOnFunction()
/netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Scalar/
H A DGVN.h48 class LoadInst; variable
311 bool processLoad(LoadInst *L);
312 bool processNonLocalLoad(LoadInst *L);
318 bool AnalyzeLoadAvailability(LoadInst *Load, MemDepResult DepInfo,
324 void AnalyzeLoadAvailability(LoadInst *Load, LoadDepVect &Deps,
328 bool PerformLoadPRE(LoadInst *Load, AvailValInBlkVect &ValuesPerBlock,
334 bool performLoopLoadPRE(LoadInst *Load, AvailValInBlkVect &ValuesPerBlock,
340 LoadInst *Load, AvailValInBlkVect &ValuesPerBlock,
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/
H A DSSAUpdater.cpp327 if (const LoadInst *LI = dyn_cast<LoadInst>(Insts[0])) in LoadAndStorePromoter()
349 SmallVector<LoadInst *, 32> LiveInLoads; in run()
368 LiveInLoads.push_back(cast<LoadInst>(User)); in run()
387 LiveInLoads.push_back(cast<LoadInst>(I)); in run()
399 if (LoadInst *L = dyn_cast<LoadInst>(&I)) { in run()
433 for (LoadInst *ALoad : LiveInLoads) { in run()
466 replaceLoadWithValue(cast<LoadInst>(User), NewVal); in run()
H A DPromoteMemoryToRegister.cpp67 if (const LoadInst *LI = dyn_cast<LoadInst>(U)) { in isAllocaPromotable()
140 LoadInst *LI = cast<LoadInst>(User); in AnalyzeAlloca()
189 return (isa<LoadInst>(I) && isa<AllocaInst>(I->getOperand(0))) || in isInterestingInstruction()
301 static void addAssumeNonNull(AssumptionCache *AC, LoadInst *LI) { in addAssumeNonNull()
318 if (isa<LoadInst>(I) || isa<StoreInst>(I)) in removeIntrinsicUsers()
369 LoadInst *LI = cast<LoadInst>(UserInst); in rewriteSingleStoreAlloca()
480 LoadInst *LI = dyn_cast<LoadInst>(U); in promoteSingleBlockAlloca()
818 if (LoadInst *LI = dyn_cast<LoadInst>(I)) in ComputeLiveInBlocks()
947 if (LoadInst *LI = dyn_cast<LoadInst>(I)) { in RenamePass()
H A DGlobalStatus.cpp87 if (const LoadInst *LI = dyn_cast<LoadInst>(I)) { in analyzeGlobalAux()
122 } else if (isa<LoadInst>(StoredVal) && in analyzeGlobalAux()
123 cast<LoadInst>(StoredVal)->getOperand(0) == GV) { in analyzeGlobalAux()
H A DRelLookupTableConverter.cpp44 LoadInst *Load = dyn_cast<LoadInst>(GEP->use_begin()->getUser()); in shouldConvertToRelLookupTable()
130 LoadInst *Load = cast<LoadInst>(GEP->use_begin()->getUser()); in convertToRelLookupTable()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Analysis/
H A DMemDerefPrinter.cpp58 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in runOnFunction()
93 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in run()
H A DMemoryDependenceAnalysis.cpp118 if (const LoadInst *LI = dyn_cast<LoadInst>(Inst)) { in GetLocation()
252 if (auto *LI = dyn_cast<LoadInst>(QueryInst)) { in getPointerDependencyFrom()
283 MemoryDependenceResults::getInvariantGroupPointerDependency(LoadInst *LI, in getInvariantGroupPointerDependency()
346 if ((isa<LoadInst>(U) || in getInvariantGroupPointerDependency()
412 LoadInst *LI = dyn_cast<LoadInst>(QueryInst); in getSimplePointerDependencyFrom()
420 if (auto *LI = dyn_cast<LoadInst>(I)) in getSimplePointerDependencyFrom()
430 return !isa<LoadInst>(I) && !isa<StoreInst>(I) && I->mayReadOrWriteMemory(); in getSimplePointerDependencyFrom()
484 if (LoadInst *LI = dyn_cast<LoadInst>(Inst)) { in getSimplePointerDependencyFrom()
845 bool isLoad = isa<LoadInst>(QueryInst); in getNonLocalPointerDependency()
872 if (LoadInst *LI = dyn_cast<LoadInst>(Inst)) { in getNonLocalPointerDependency()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Analysis/
H A DLoads.h25 class LoadInst; variable
83 bool isDereferenceableAndAlignedInLoop(LoadInst *LI, Loop *L,
129 Value *FindAvailableLoadedValue(LoadInst *Load,
141 Value *FindAvailableLoadedValue(LoadInst *Load, AAResults &AA, bool *IsLoadCSE,
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/
H A DLowerAtomic.cpp30 LoadInst *Orig = Builder.CreateLoad(Val->getType(), Ptr); in LowerAtomicCmpXchgInst()
48 LoadInst *Orig = Builder.CreateLoad(Val->getType(), Ptr); in lowerAtomicRMWInst()
108 static bool LowerLoadInst(LoadInst *LI) { in LowerLoadInst()
127 else if (LoadInst *LI = dyn_cast<LoadInst>(&Inst)) { in runOnBasicBlock()
H A DLoopLoadElimination.cpp90 LoadInst *Load;
93 StoreToLoadForwardingCandidate(LoadInst *Load, StoreInst *Store) in StoreToLoadForwardingCandidate()
157 static bool isLoadConditional(LoadInst *Load, Loop *L) { in isLoadConditional()
195 if (isa<LoadInst>(Source)) in findStoreToLoadDependences()
197 if (isa<LoadInst>(Destination)) in findStoreToLoadDependences()
213 auto *Load = dyn_cast<LoadInst>(Destination); in findStoreToLoadDependences()
263 DenseMap<LoadInst *, const StoreToLoadForwardingCandidate *>; in removeDependencesFromMultipleStores()
344 LoadInst *LastLoad = in findPointersWrittenOnForwardingPath()
436 Value *Initial = new LoadInst( in propagateStoredValueToLoadUsers()
H A DSROA.cpp776 void visitLoadInst(LoadInst &LI) { in visitLoadInst()
975 if (LoadInst *LI = dyn_cast<LoadInst>(I)) { in hasUnsafePHIOrSelectUse()
1148 if (LoadInst *LI = dyn_cast<LoadInst>(U->getUser())) { in findCommonType()
1211 LoadInst *LI = dyn_cast<LoadInst>(U); in isSafePHIToSpeculate()
1269 LoadInst *SomeLoad = cast<LoadInst>(PN.user_back()); in speculatePHINodeLoads()
1283 LoadInst *LI = cast<LoadInst>(PN.user_back()); in speculatePHINodeLoads()
1306 LoadInst *Load = PredBuilder.CreateAlignedLoad( in speculatePHINodeLoads()
1339 LoadInst *LI = dyn_cast<LoadInst>(U); in isSafeSelectToSpeculate()
1365 LoadInst *LI = cast<LoadInst>(SI.user_back()); in speculateSelectInstLoads()
1369 LoadInst *TL = IRB.CreateLoad(LI->getType(), TV, in speculateSelectInstLoads()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/InstCombine/
H A DInstCombineLoadStoreAlloca.cpp60 if (auto *LI = dyn_cast<LoadInst>(I)) { in isOnlyCopiedFromConstantMemory()
257 if (LoadInst *Load = dyn_cast<LoadInst>(Inst)) { in collectUsers()
282 if (auto *LT = dyn_cast<LoadInst>(I)) { in replace()
285 auto *NewI = new LoadInst(LT->getType(), V, "", LT->isVolatile(), in replace()
459 LoadInst *InstCombinerImpl::combineLoadToNewType(LoadInst &LI, Type *NewTy, in combineLoadToNewType()
472 LoadInst *NewLoad = Builder.CreateAlignedLoad( in combineLoadToNewType()
577 LoadInst &LI) { in combineLoadToOperationType()
609 LoadInst *NewLoad = IC.combineLoadToNewType(LI, CI->getDestTy()); in combineLoadToOperationType()
621 static Instruction *unpackLoadToAggregate(InstCombinerImpl &IC, LoadInst &LI) { in unpackLoadToAggregate()
638 LoadInst *NewLoad = IC.combineLoadToNewType(LI, ST->getTypeAtIndex(0U), in unpackLoadToAggregate()
[all …]
H A DInstCombinePHI.cpp119 if (LoadInst *LoadI = dyn_cast<LoadInst>(U)) { in foldIntegerTypedPHI()
174 auto *LoadI = dyn_cast<LoadInst>(Arg); in foldIntegerTypedPHI()
263 LoadInst *LoadI = dyn_cast<LoadInst>(IncomingVal); in foldIntegerTypedPHI()
591 static bool isSafeAndProfitableToSinkLoad(LoadInst *L) { in isSafeAndProfitableToSinkLoad()
609 if (isa<LoadInst>(U)) continue; in isSafeAndProfitableToSinkLoad()
636 LoadInst *FirstLI = cast<LoadInst>(PN.getIncomingValue(0)); in foldPHIArgLoadIntoPHI()
667 LoadInst *LI = dyn_cast<LoadInst>(PN.getIncomingValue(i)); in foldPHIArgLoadIntoPHI()
697 LoadInst *NewLI = in foldPHIArgLoadIntoPHI()
698 new LoadInst(FirstLI->getType(), NewPN, "", isVolatile, LoadAlignment); in foldPHIArgLoadIntoPHI()
718 LoadInst *LI = cast<LoadInst>(PN.getIncomingValue(i)); in foldPHIArgLoadIntoPHI()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/
H A DInterleavedLoadCombinePass.cpp94 LoadInst *findFirstLoad(const std::set<LoadInst *> &LIs);
651 LoadInst *LI;
653 ElementInfo(Polynomial Offset = Polynomial(), LoadInst *LI = nullptr) in ElementInfo()
664 std::set<LoadInst *> LIs;
716 LoadInst *LI = dyn_cast<LoadInst>(V); in compute()
868 static bool computeFromLI(LoadInst *LI, VectorInfo &Result, in computeFromLI()
1101 LoadInst *
1102 InterleavedLoadCombineImpl::findFirstLoad(const std::set<LoadInst *> &LIs) { in findFirstLoad()
1111 return cast<LoadInst>(FLI); in findFirstLoad()
1121 LoadInst *InsertionPoint = InterleavedLoad.front().EI[0].LI; in combine()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Instrumentation/
H A DThreadSanitizer.cpp422 } else if (LoadInst *L = dyn_cast<LoadInst>(Addr)) { in addrPointsToConstantData()
452 : cast<LoadInst>(I)->getPointerOperand(); in chooseInstructionsToInstrument()
464 ClDistinguishVolatile && (cast<LoadInst>(I)->isVolatile() || in chooseInstructionsToInstrument()
503 if (LoadInst *LI = dyn_cast<LoadInst>(I)) in isAtomic()
551 else if (isa<LoadInst>(Inst) || isa<StoreInst>(Inst)) in sanitizeFunction()
616 : cast<LoadInst>(II.Inst)->getPointerOperand(); in instrumentLoadOrStore()
653 : cast<LoadInst>(II.Inst)->getAlignment(); in instrumentLoadOrStore()
658 : cast<LoadInst>(II.Inst)->isVolatile()); in instrumentLoadOrStore()
743 if (LoadInst *LI = dyn_cast<LoadInst>(I)) { in instrumentAtomic()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Vectorize/
H A DLoadStoreVectorizer.cpp296 if (LoadInst *L = dyn_cast<LoadInst>(I)) in getPointerAddressSpace()
652 bool IsLoadChain = isa<LoadInst>(Chain[0]); in getVectorizablePrefix()
656 assert(isa<LoadInst>(I) && in getVectorizablePrefix()
665 if (isa<LoadInst>(I) || isa<StoreInst>(I)) { in getVectorizablePrefix()
711 auto *MemLoad = dyn_cast<LoadInst>(MemInstr); in getVectorizablePrefix()
712 auto *ChainLoad = dyn_cast<LoadInst>(ChainInstr); in getVectorizablePrefix()
718 auto IsInvariantLoad = [](const LoadInst *LI) -> bool { in getVectorizablePrefix()
800 if (LoadInst *LI = dyn_cast<LoadInst>(&I)) { in collectInstructions()
978 if (isa<LoadInst>(*Operands.begin())) in vectorizeInstructions()
1145 LoadInst *L0 = cast<LoadInst>(Chain[0]); in vectorizeLoadChain()
[all …]

1234567891011