Home
last modified time | relevance | path

Searched refs:StoreInst (Results 1 – 25 of 217) sorted by relevance

123456789

/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/
H A DMergedLoadStoreMotion.cpp118 StoreInst *canSinkFromBlock(BasicBlock *BB, StoreInst *SI);
119 PHINode *getPHIOperand(BasicBlock *BB, StoreInst *S0, StoreInst *S1);
122 bool canSinkStoresAndGEPs(StoreInst *S0, StoreInst *S1) const;
123 void sinkStoresAndGEPs(BasicBlock *BB, StoreInst *SinkCand,
124 StoreInst *ElseInst);
187 StoreInst *MergedLoadStoreMotion::canSinkFromBlock(BasicBlock *BB1, in canSinkFromBlock()
188 StoreInst *Store0) { in canSinkFromBlock()
192 auto *Store1 = dyn_cast<StoreInst>(&Inst); in canSinkFromBlock()
210 PHINode *MergedLoadStoreMotion::getPHIOperand(BasicBlock *BB, StoreInst *S0, in getPHIOperand()
211 StoreInst *S1) { in getPHIOperand()
[all …]
H A DLowerAtomic.cpp113 static bool LowerStoreInst(StoreInst *SI) { in LowerStoreInst()
130 } else if (StoreInst *SI = dyn_cast<StoreInst>(&Inst)) { in runOnBasicBlock()
H A DDeadStoreElimination.cpp168 if (isa<StoreInst>(I)) in hasAnalyzableMemoryWrite()
209 if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) in getLocForWrite()
241 if (StoreInst *SI = dyn_cast<StoreInst>(I)) in isRemovable()
279 if (isa<StoreInst>(I)) in isShortenableAtTheEnd()
742 StoreInst *Earlier, StoreInst *Later, int64_t InstWriteOffset, in tryToMergePartialOverlappingStores()
1236 if (auto SI = dyn_cast<StoreInst>(UseInst)) in isReadClobber()
1713 if (auto *SI = dyn_cast<StoreInst>(NI)) in isDSEBarrier()
1768 StoreInst *Store = dyn_cast<StoreInst>(Def->getMemoryInst()); in storeIsNoop()
1951 auto *Earlier = dyn_cast<StoreInst>(NI); in eliminateDeadStores()
1952 auto *Later = dyn_cast<StoreInst>(SI); in eliminateDeadStores()
[all …]
H A DSROA.cpp794 void visitStoreInst(StoreInst &SI) { in visitStoreInst()
980 if (StoreInst *SI = dyn_cast<StoreInst>(I)) { in hasUnsafePHIOrSelectUse()
1150 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in findCommonType()
1866 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in isVectorPromotionViableForSlice()
1922 else if (auto *SI = dyn_cast<StoreInst>(S.getUse()->getUser())) in isVectorPromotionViable()
2050 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in isIntegerWideningViableForSlice()
2613 bool rewriteVectorizedStoreInst(Value *V, StoreInst &SI, Value *OldOp, in rewriteVectorizedStoreInst()
2633 StoreInst *Store = IRB.CreateAlignedStore(V, &NewAI, NewAI.getAlign()); in rewriteVectorizedStoreInst()
2642 bool rewriteIntegerStore(Value *V, StoreInst &SI, AAMDNodes AATags) { in rewriteIntegerStore()
2655 StoreInst *Store = IRB.CreateAlignedStore(V, &NewAI, NewAI.getAlign()); in rewriteIntegerStore()
[all …]
H A DLoopDataPrefetch.cpp254 Writes = isa<StoreInst>(I); in addInstruction()
264 if (isa<StoreInst>(I) && PtrDiff == 0) in addInstruction()
326 } else if (StoreInst *SMemI = dyn_cast<StoreInst>(&I)) { in runOnLoop()
H A DMemCpyOptimizer.cpp122 if (!isa<StoreInst>(SI)) in isProfitableToUseMemset()
174 if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) in addInst()
180 void addStore(int64_t OffsetFromFirst, StoreInst *SI) { in addStore()
409 if (!isa<StoreInst>(BI) && !isa<MemSetInst>(BI)) { in tryMergingIntoMemset()
418 if (StoreInst *NextStore = dyn_cast<StoreInst>(BI)) { in tryMergingIntoMemset()
525 bool MemCpyOptPass::moveUp(StoreInst *SI, Instruction *P, const LoadInst *LI) { in moveUp()
587 } else if (isa<LoadInst>(C) || isa<StoreInst>(C) || isa<VAArgInst>(C)) { in moveUp()
649 bool MemCpyOptPass::processStore(StoreInst *SI, BasicBlock::iterator &BBI) { in processStore()
1684 if (StoreInst *SI = dyn_cast<StoreInst>(I)) in iterateOnFunction()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/
H A DX86AvoidStoreForwardingBlocks.cpp102 void breakBlockedCopies(MachineInstr *LoadInst, MachineInstr *StoreInst,
106 MachineInstr *StoreInst, int64_t StDispImm,
110 MachineInstr *StoreInst, unsigned NStoreOpcode,
386 int64_t LoadDisp, MachineInstr *StoreInst, in buildCopy() argument
391 MachineOperand &StoreBase = getBaseOperand(StoreInst); in buildCopy()
394 MachineMemOperand *SMMO = *StoreInst->memoperands_begin(); in buildCopy()
413 MachineInstr *StInst = StoreInst; in buildCopy()
414 auto PrevInstrIt = prev_nodbg(MachineBasicBlock::instr_iterator(StoreInst), in buildCopy()
430 MachineOperand &StoreSrcVReg = StoreInst->getOperand(X86::AddrNumOperands); in buildCopy()
437 int64_t LdDispImm, MachineInstr *StoreInst, in buildCopies() argument
[all …]
H A DX86PreAMXConfig.cpp164 new StoreInst(PaletteValue, PalettePos, Pos); in preWriteTileCfg()
178 new StoreInst(Row, RowPos, Pos); in preWriteTileCfg()
179 new StoreInst(Col, ColPos, Pos); in preWriteTileCfg()
205 Instruction *Init0 = new StoreInst(Val0, Addr, false, Alignment, Cfg); in addTileConfig()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/
H A DPromoteMemoryToRegister.cpp72 } else if (const StoreInst *SI = dyn_cast<StoreInst>(U)) { in isAllocaPromotable()
109 StoreInst *OnlyStore;
135 if (StoreInst *SI = dyn_cast<StoreInst>(User)) { in AnalyzeAlloca()
190 (isa<StoreInst>(I) && isa<AllocaInst>(I->getOperand(1))); in isInterestingInstruction()
318 if (isa<LoadInst>(I) || isa<StoreInst>(I)) in removeIntrinsicUsers()
357 StoreInst *OnlyStore = Info.OnlyStore; in rewriteSingleStoreAlloca()
466 using StoresByIndexTy = SmallVector<std::pair<unsigned, StoreInst *>, 64>; in promoteSingleBlockAlloca()
470 if (StoreInst *SI = dyn_cast<StoreInst>(U)) in promoteSingleBlockAlloca()
489 std::make_pair(LoadIdx, static_cast<StoreInst *>(nullptr)), in promoteSingleBlockAlloca()
522 StoreInst *SI = cast<StoreInst>(AI->user_back()); in promoteSingleBlockAlloca()
[all …]
H A DSSAUpdater.cpp330 SomeVal = cast<StoreInst>(Insts[0])->getOperand(0); in LoadAndStorePromoter()
363 if (StoreInst *SI = dyn_cast<StoreInst>(User)) { in run()
376 if (isa<StoreInst>(I)) { in run()
415 if (StoreInst *SI = dyn_cast<StoreInst>(&I)) { in run()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/
H A DAMDGPURewriteOutArguments.cpp122 StoreInst *SI = dyn_cast<StoreInst>(U.getUser()); in checkArgumentUses()
160 U.getOperandNo() != StoreInst::getPointerOperandIndex()) in checkArgumentUses()
269 SmallVector<std::pair<ReturnInst *, StoreInst *>, 4> ReplaceableStores; in runOnFunction()
289 StoreInst *SI = nullptr; in runOnFunction()
291 SI = dyn_cast<StoreInst>(Q.getInst()); in runOnFunction()
305 for (std::pair<ReturnInst *, StoreInst *> Store : ReplaceableStores) { in runOnFunction()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/InstCombine/
H A DInstCombineLoadStoreAlloca.cpp482 static StoreInst *combineStoreToNewValue(InstCombinerImpl &IC, StoreInst &SI, in combineStoreToNewValue()
492 StoreInst *NewStore = IC.Builder.CreateAlignedStore( in combineStoreToNewValue()
912 static bool canSimplifyNullStoreOrGEP(StoreInst &SI) { in canSimplifyNullStoreOrGEP()
984 StoreInst *SI = new StoreInst(UndefValue::get(LI.getType()), in visitLoadInst()
1113 static bool combineStoreToValueType(InstCombinerImpl &IC, StoreInst &SI) { in combineStoreToValueType()
1151 static bool unpackStoreToAggregate(InstCombinerImpl &IC, StoreInst &SI) { in unpackStoreToAggregate()
1292 StoreInst &SI) { in removeBitcastsFromLoadStoreOnMinMax()
1319 auto *SI = dyn_cast<StoreInst>(U); in removeBitcastsFromLoadStoreOnMinMax()
1331 auto *USI = cast<StoreInst>(UI); in removeBitcastsFromLoadStoreOnMinMax()
1340 Instruction *InstCombinerImpl::visitStoreInst(StoreInst &SI) { in visitStoreInst()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Instrumentation/
H A DThreadSanitizer.cpp450 const bool IsWrite = isa<StoreInst>(*I); in chooseInstructionsToInstrument()
451 Value *Addr = IsWrite ? cast<StoreInst>(I)->getPointerOperand() in chooseInstructionsToInstrument()
465 cast<StoreInst>(WI.Inst)->isVolatile()); in chooseInstructionsToInstrument()
505 if (StoreInst *SI = dyn_cast<StoreInst>(I)) in isAtomic()
551 else if (isa<LoadInst>(Inst) || isa<StoreInst>(Inst)) in sanitizeFunction()
614 const bool IsWrite = isa<StoreInst>(*II.Inst); in instrumentLoadOrStore()
615 Value *Addr = IsWrite ? cast<StoreInst>(II.Inst)->getPointerOperand() in instrumentLoadOrStore()
629 Value *StoredValue = cast<StoreInst>(II.Inst)->getValueOperand(); in instrumentLoadOrStore()
652 const unsigned Alignment = IsWrite ? cast<StoreInst>(II.Inst)->getAlignment() in instrumentLoadOrStore()
657 (IsWrite ? cast<StoreInst>(II.Inst)->isVolatile() in instrumentLoadOrStore()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Scalar/
H A DMemCpyOptimizer.h38 class StoreInst; variable
63 bool processStore(StoreInst *SI, BasicBlock::iterator &BBI);
76 bool moveUp(StoreInst *SI, Instruction *P, const LoadInst *LI);
/netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/
H A DGCRootLowering.cpp128 if (isa<AllocaInst>(I) || isa<GetElementPtrInst>(I) || isa<StoreInst>(I) || in CouldBecomeSafePoint()
151 if (StoreInst *SI = dyn_cast<StoreInst>(IP)) in InsertRootInitializers()
161 new StoreInst( in InsertRootInitializers()
205 Value *St = new StoreInst(CI->getArgOperand(0), in DoLowering()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Vectorize/
H A DLoadStoreVectorizer.cpp298 if (StoreInst *S = dyn_cast<StoreInst>(I)) in getPointerAddressSpace()
659 assert(isa<StoreInst>(I) && in getVectorizablePrefix()
665 if (isa<LoadInst>(I) || isa<StoreInst>(I)) { in getVectorizablePrefix()
726 if (isa<StoreInst>(MemInstr) && ChainLoad && in getVectorizablePrefix()
731 if (MemLoad && isa<StoreInst>(ChainInstr) && in getVectorizablePrefix()
847 } else if (StoreInst *SI = dyn_cast<StoreInst>(&I)) { in collectInstructions()
992 StoreInst *S0 = cast<StoreInst>(Chain[0]); in vectorizeStoreChain()
997 StoreTy = cast<StoreInst>(I)->getValueOperand()->getType(); in vectorizeStoreChain()
1103 StoreInst *Store = cast<StoreInst>(Chain[I]); in vectorizeStoreChain()
1118 StoreInst *Store = cast<StoreInst>(Chain[I]); in vectorizeStoreChain()
[all …]
H A DVPlanTransforms.cpp66 } else if (StoreInst *Store = dyn_cast<StoreInst>(Inst)) { in VPInstructionsToVPRecipes()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/IPO/
H A DGlobalOpt.cpp214 if (StoreInst *SI = dyn_cast<StoreInst>(U)) { in CleanupPointerRootUsers()
306 } else if (StoreInst *SI = dyn_cast<StoreInst>(U)) { in CleanupConstantGlobalUsers()
420 if (StoreInst *SI = dyn_cast<StoreInst>(I)) in isSafeSROAElementUse()
641 if (auto *Store = dyn_cast<StoreInst>(U)) { in SRAGlobal()
678 } else if (const StoreInst *SI = dyn_cast<StoreInst>(U)) { in AllUsesOfValueWillTrapIfNull()
719 } else if (isa<StoreInst>(U)) { in AllUsesOfLoadedValueWillTrapIfNull()
740 } else if (StoreInst *SI = dyn_cast<StoreInst>(I)) { in OptimizeAwayTrappingUsesOfValue()
821 } else if (isa<StoreInst>(GlobalUser)) { in OptimizeAwayTrappingUsesOfLoads()
942 if (StoreInst *SI = dyn_cast<StoreInst>(GV->user_back())) { in OptimizeGlobalAddressOfMalloc()
944 new StoreInst(ConstantInt::getTrue(GV->getContext()), InitBool, false, in OptimizeGlobalAddressOfMalloc()
[all …]
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/NVPTX/
H A DNVPTXLowerAggrCopies.cpp79 if (StoreInst *SI = dyn_cast<StoreInst>(LI->user_back())) { in runOnFunction()
106 auto *SI = cast<StoreInst>(*LI->user_begin()); in runOnFunction()
/netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Vectorize/
H A DSLPVectorizer.h45 class StoreInst; variable
59 using StoreList = SmallVector<StoreInst *, 8>;
138 bool vectorizeStores(ArrayRef<StoreInst *> Stores, slpvectorizer::BoUpSLP &R);
/netbsd-src/external/apache2/llvm/dist/llvm/lib/IR/
H A DInstruction.cpp420 if (const StoreInst *SI = dyn_cast<StoreInst>(I1)) in haveSameSpecialState()
421 return SI->isVolatile() == cast<StoreInst>(I2)->isVolatile() && in haveSameSpecialState()
422 (SI->getAlignment() == cast<StoreInst>(I2)->getAlignment() || in haveSameSpecialState()
424 SI->getOrdering() == cast<StoreInst>(I2)->getOrdering() && in haveSameSpecialState()
425 SI->getSyncScopeID() == cast<StoreInst>(I2)->getSyncScopeID(); in haveSameSpecialState()
559 return !cast<StoreInst>(this)->isUnordered(); in mayReadFromMemory()
594 return cast<StoreInst>(this)->getOrdering() != AtomicOrdering::NotAtomic; in isAtomic()
629 return cast<StoreInst>(this)->isVolatile(); in isVolatile()
/netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Utils/
H A DAutoInitRemark.h30 class StoreInst; variable
45 void inspectStore(StoreInst &SI);
H A DVNCoercion.h26 class StoreInst; variable
55 StoreInst *DepSI, const DataLayout &DL);
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/ObjCARC/
H A DObjCARCContract.cpp197 static StoreInst *findSafeStoreForStoreStrongContraction(LoadInst *Load, in findSafeStoreForStoreStrongContraction()
201 StoreInst *Store = nullptr; in findSafeStoreForStoreStrongContraction()
256 Store = dyn_cast<StoreInst>(Inst); in findSafeStoreForStoreStrongContraction()
284 findRetainForStoreStrongContraction(Value *New, StoreInst *Store, in findRetainForStoreStrongContraction()
350 StoreInst *Store = in tryToContractReleaseIntoStoreStrong()
481 new StoreInst(Null, CI->getArgOperand(0), CI); in tryToPeepholeInstruction()
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Analysis/
H A DLoads.cpp385 } else if (StoreInst *SI = dyn_cast<StoreInst>(BBI)) { in isSafeToLoadUnconditionally()
497 if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) { in getAvailableLoadStore()
555 if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) { in findAvailablePtrLoadStore()

123456789