| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Utils/ |
| H A D | SimplifyLibCalls.h | 23 class CallInst; variable 51 Value *optimizeCall(CallInst *CI, IRBuilderBase &B); 54 Value *optimizeMemCpyChk(CallInst *CI, IRBuilderBase &B); 55 Value *optimizeMemMoveChk(CallInst *CI, IRBuilderBase &B); 56 Value *optimizeMemSetChk(CallInst *CI, IRBuilderBase &B); 59 Value *optimizeStrpCpyChk(CallInst *CI, IRBuilderBase &B, LibFunc Func); 60 Value *optimizeStrpNCpyChk(CallInst *CI, IRBuilderBase &B, LibFunc Func); 61 Value *optimizeStrLenChk(CallInst *CI, IRBuilderBase &B); 62 Value *optimizeMemPCpyChk(CallInst *CI, IRBuilderBase &B); 63 Value *optimizeMemCCpyChk(CallInst *CI, IRBuilderBase &B); [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/ |
| H A D | LibCallsShrinkWrap.cpp | 78 void visitCallInst(CallInst &CI) { checkCandidate(CI); } in visitCallInst() 93 bool perform(CallInst *CI); 94 void checkCandidate(CallInst &CI); 95 void shrinkWrapCI(CallInst *CI, Value *Cond); 96 bool performCallDomainErrorOnly(CallInst *CI, const LibFunc &Func); 97 bool performCallErrors(CallInst *CI, const LibFunc &Func); 98 bool performCallRangeErrorOnly(CallInst *CI, const LibFunc &Func); 99 Value *generateOneRangeCond(CallInst *CI, const LibFunc &Func); 100 Value *generateTwoRangeCond(CallInst *CI, const LibFunc &Func); 101 Value *generateCondForPow(CallInst *CI, const LibFunc &Func); [all …]
|
| H A D | SimplifyLibCalls.cpp | 71 static bool callHasFloatingPointArgument(const CallInst *CI) { in callHasFloatingPointArgument() 77 static bool callHasFP128Argument(const CallInst *CI) { in callHasFP128Argument() 83 static Value *convertStrToNumber(CallInst *CI, StringRef &Str, int64_t Base) { in convertStrToNumber() 120 static bool canTransformToMemCmp(CallInst *CI, Value *Str, uint64_t Len, in canTransformToMemCmp() 134 static void annotateDereferenceableBytes(CallInst *CI, in annotateDereferenceableBytes() 161 static void annotateNonNullNoUndefBasedOnAccess(CallInst *CI, in annotateNonNullNoUndefBasedOnAccess() 182 static void annotateNonNullAndDereferenceable(CallInst *CI, ArrayRef<unsigned> ArgNos, in annotateNonNullAndDereferenceable() 202 Value *LibCallSimplifier::optimizeStrCat(CallInst *CI, IRBuilderBase &B) { in optimizeStrCat() 244 Value *LibCallSimplifier::optimizeStrNCat(CallInst *CI, IRBuilderBase &B) { in optimizeStrNCat() 287 Value *LibCallSimplifier::optimizeStrChr(CallInst *CI, IRBuilderBase &B) { in optimizeStrChr() [all …]
|
| H A D | EscapeEnumerator.cpp | 44 if (CallInst *CI = CurBB->getTerminatingMustTailCall()) in Next() 63 if (CallInst *CI = dyn_cast<CallInst>(&II)) in Next() 92 CallInst *CI = cast<CallInst>(Calls[--I]); in Next()
|
| H A D | EntryExitInstrumenter.cpp | 38 CallInst *Call = CallInst::Create(Fn, "", InsertionPt); in insertCall() 49 Instruction *RetAddr = CallInst::Create( in insertCall() 58 CallInst *Call = in insertCall() 59 CallInst::Create(Fn, ArrayRef<Value *>(Args), "", InsertionPt); in insertCall() 102 if (CallInst *CI = BB.getTerminatingMustTailCall()) in runOnFunction()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/IR/ |
| H A D | IRBuilder.cpp | 72 static CallInst *createCallHelper(Function *Callee, ArrayRef<Value *> Ops, in createCallHelper() 77 CallInst *CI = Builder->CreateCall(Callee, Ops, OpBundles, Name); 90 CallInst *CI = createCallHelper(TheFn, {}, this, Name); in CreateVScale() 113 CallInst *IRBuilderBase::CreateMemSet(Value *Ptr, Value *Val, Value *Size, in CreateMemSet() 123 CallInst *CI = createCallHelper(TheFn, Ops, this); in CreateMemSet() 141 CallInst *IRBuilderBase::CreateElementUnorderedAtomicMemSet( in CreateElementUnorderedAtomicMemSet() 152 CallInst *CI = createCallHelper(TheFn, Ops, this); in CreateElementUnorderedAtomicMemSet() 169 CallInst *IRBuilderBase::CreateMemTransferInst( in CreateMemTransferInst() 181 CallInst *CI = createCallHelper(TheFn, Ops, this); in CreateMemTransferInst() 206 CallInst *IRBuilderBase::CreateMemCpyInline(Value *Dst, MaybeAlign DstAlign, in CreateMemCpyInline() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/SelectionDAG/ |
| H A D | SelectionDAGBuilder.h | 46 class CallInst; variable 631 void LowerDeoptimizeCall(const CallInst *CI); 738 void visitMaskedLoad(const CallInst &I, bool IsExpanding = false); 739 void visitMaskedStore(const CallInst &I, bool IsCompressing = false); 740 void visitMaskedGather(const CallInst &I); 741 void visitMaskedScatter(const CallInst &I); 746 void visitCall(const CallInst &I); 747 bool visitMemCmpBCmpCall(const CallInst &I); 748 bool visitMemPCpyCall(const CallInst &I); 749 bool visitMemChrCall(const CallInst &I); [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPULibCalls.cpp | 61 bool replaceWithNative(CallInst *CI, const FuncInfo &FInfo); 66 bool TDOFold(CallInst *CI, const FuncInfo &FInfo); 71 bool fold_recip(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 74 bool fold_divide(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 77 bool fold_pow(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 80 bool fold_rootn(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 83 bool fold_fma_mad(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); 86 bool sincosUseNative(CallInst *aCI, const FuncInfo &FInfo); 91 bool evaluateCall(CallInst *aCI, FuncInfo &FInfo); 94 bool fold_exp(CallInst *CI, IRBuilder<> &B, const FuncInfo &FInfo); [all …]
|
| H A D | SIAnnotateControlFlow.cpp | 188 if (const CallInst *CI = dyn_cast<CallInst>(&I)) in hasKill() 207 Value *Ret = CallInst::Create(If, Term->getCondition(), "", Term); in openIf() 217 Value *Ret = CallInst::Create(Else, popSaved(), "", Term); in insertElse() 235 return CallInst::Create(IfBreak, Args, "", Insert); in handleLoopCondition() 244 return CallInst::Create(IfBreak, Args, "", Insert); in handleLoopCondition() 279 Term->setCondition(CallInst::Create(Loop, Arg, "", Term)); in handleLoop() 316 CallInst::Create(EndCf, Exec, "", FirstInsertionPt); in closeControlFlow()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/ObjCARC/ |
| H A D | ObjCARC.h | 41 Value *OldArg = cast<CallInst>(CI)->getArgOperand(0); in EraseInstruction() 68 if (const auto *C = dyn_cast<CallInst>(Opnd)) in getreturnRVOperand() 101 CallInst *createCallInstWithColors( 118 CallInst *insertRVCall(Instruction *InsertPt, CallBase *AnnotatedCall); 121 CallInst *insertRVCallWithColors( 127 if (auto *CI = dyn_cast<CallInst>(I)) in contains() 133 void eraseInst(CallInst *CI) { in eraseInst() 138 if (auto *CI = dyn_cast<CallInst>(*U)) in eraseInst() 156 DenseMap<CallInst *, CallBase *> RVCalls;
|
| H A D | ObjCARCOpts.cpp | 103 cast<CallInst>(Arg)->getArgOperand(0)); in FindSingleUseIdentifiedObject() 656 cast<CallInst>(RetainRV)->setCalledFunction(NewDecl); in OptimizeRetainRVCall() 695 cast<CallInst>(AutoreleaseRV)->getArgOperand(0)); in OptimizeInlinedAutoreleaseRVCall() 700 Inst->replaceAllUsesWith(cast<CallInst>(Inst)->getArgOperand(0)); in OptimizeInlinedAutoreleaseRVCall() 708 Value *CallArg = cast<CallInst>(Inst)->getArgOperand(0); in OptimizeInlinedAutoreleaseRVCall() 709 CallInst *Release = CallInst::Create( in OptimizeInlinedAutoreleaseRVCall() 763 CallInst *AutoreleaseRVCI = cast<CallInst>(AutoreleaseRV); in OptimizeAutoreleaseRVCall() 774 CloneCallInstForBB(CallInst &CI, BasicBlock &BB, in CloneCallInstForBB() 793 return CallInst::Create(&CI, OpBundles); in CloneCallInstForBB() 853 if (auto *CI = dyn_cast<CallInst>(Inst)) in OptimizeIndividualCalls() [all …]
|
| H A D | ObjCARC.cpp | 46 CallInst *objcarc::createCallInstWithColors( in createCallInstWithColors() 62 return CallInst::Create(FTy, Callee, Args, OpBundles, NameStr, InsertBefore); in createCallInstWithColors() 96 CallInst *BundledRetainClaimRVs::insertRVCall(Instruction *InsertPt, in insertRVCall() 102 CallInst *BundledRetainClaimRVs::insertRVCallWithColors( in insertRVCallWithColors() 124 if (auto *CI = dyn_cast<CallInst>(P.second)) in ~BundledRetainClaimRVs() 125 CI->setTailCallKind(CallInst::TCK_NoTail); in ~BundledRetainClaimRVs()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Analysis/ |
| H A D | MemoryBuiltins.h | 32 class CallInst; variable 125 const CallInst * 128 inline CallInst * 131 return const_cast<CallInst *>(extractMallocCall((const Value *)I, GetTLI)); in extractMallocCall() 139 PointerType *getMallocType(const CallInst *CI, const TargetLibraryInfo *TLI); 146 Type *getMallocAllocatedType(const CallInst *CI, const TargetLibraryInfo *TLI); 153 Value *getMallocArraySize(CallInst *CI, const DataLayout &DL, 163 const CallInst *extractCallocCall(const Value *I, const TargetLibraryInfo *TLI); 164 inline CallInst *extractCallocCall(Value *I, const TargetLibraryInfo *TLI) { in extractCallocCall() 165 return const_cast<CallInst*>(extractCallocCall((const Value*)I, TLI)); in extractCallocCall() [all …]
|
| H A D | TypeMetadataUtils.h | 23 class CallInst; variable 48 SmallVectorImpl<CallInst *> &Assumes, const CallInst *CI, 57 const CallInst *CI, DominatorTree &DT);
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
| H A D | PreISelIntrinsicLowering.cpp | 40 auto CI = dyn_cast<CallInst>(I->getUser()); in lowerLoadRelative() 63 static CallInst::TailCallKind getOverridingTailCallKind(const Function &F) { in getOverridingTailCallKind() 66 return CallInst::TCK_Tail; in getOverridingTailCallKind() 68 return CallInst::TCK_NoTail; in getOverridingTailCallKind() 69 return CallInst::TCK_None; in getOverridingTailCallKind() 91 CallInst::TailCallKind OverridingTCK = getOverridingTailCallKind(F); in lowerObjCCall() 94 auto *CI = cast<CallInst>(I->getUser()); in lowerObjCCall() 100 CallInst *NewCI = Builder.CreateCall(FCache, Args); in lowerObjCCall() 110 CallInst::TailCallKind TCK = CI->getTailCallKind(); in lowerObjCCall()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/BPF/ |
| H A D | BPFAbstractMemberAccess.cpp | 108 auto *NewInst = CallInst::Create(Fn, {SeqNumVal, Input}); in insertPassThrough() 130 typedef std::stack<std::pair<CallInst *, CallInfo>> CallInfoStack; 146 std::map<CallInst *, std::pair<CallInst *, CallInfo>> AIChain; 150 std::map<CallInst *, CallInfo> BaseAICalls; 154 void traceAICall(CallInst *Call, CallInfo &ParentInfo); 155 void traceBitCast(BitCastInst *BitCast, CallInst *Parent, 157 void traceGEP(GetElementPtrInst *GEP, CallInst *Parent, 161 bool IsPreserveDIAccessIndexCall(const CallInst *Call, CallInfo &Cinfo); 165 void replaceWithGEP(std::vector<CallInst *> &CallList, 174 Value *computeBaseAndAccessKey(CallInst *Call, CallInfo &CInfo, [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/IR/ |
| H A D | BasicBlock.h | 35 class CallInst; variable 131 const CallInst *getTerminatingDeoptimizeCall() const; 132 CallInst *getTerminatingDeoptimizeCall() { in getTerminatingDeoptimizeCall() 133 return const_cast<CallInst *>( in getTerminatingDeoptimizeCall() 140 const CallInst *getPostdominatingDeoptimizeCall() const; 141 CallInst *getPostdominatingDeoptimizeCall() { in getPostdominatingDeoptimizeCall() 142 return const_cast<CallInst *>( in getPostdominatingDeoptimizeCall() 149 const CallInst *getTerminatingMustTailCall() const; 150 CallInst *getTerminatingMustTailCall() { in getTerminatingMustTailCall() 151 return const_cast<CallInst *>( in getTerminatingMustTailCall()
|
| H A D | IRBuilder.h | 580 CallInst *CreateMemSet(Value *Ptr, Value *Val, uint64_t Size, 588 CallInst *CreateMemSet(Value *Ptr, Value *Val, Value *Size, MaybeAlign Align, 599 CallInst *CreateElementUnorderedAtomicMemSet(Value *Ptr, Value *Val, 610 CallInst *CreateElementUnorderedAtomicMemSet(Value *Ptr, Value *Val, 622 CallInst *CreateMemCpy(Value *Dst, MaybeAlign DstAlign, Value *Src, 633 CallInst *CreateMemTransferInst( 639 CallInst *CreateMemCpy(Value *Dst, MaybeAlign DstAlign, Value *Src, 650 CallInst *CreateMemCpyInline(Value *Dst, MaybeAlign DstAlign, Value *Src, 661 CallInst *CreateElementUnorderedAtomicMemCpy( 667 CallInst *CreateMemMove(Value *Dst, MaybeAlign DstAlign, Value *Src, [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/ |
| H A D | TailRecursionElimination.cpp | 233 SmallVector<CallInst *, 32> DeferredTails; in markTails() 242 CallInst *CI = dyn_cast<CallInst>(&I); in markTails() 319 for (CallInst *CI : DeferredTails) { in markTails() 338 static bool canMoveAboveCall(Instruction *I, CallInst *CI, AliasAnalysis *AA) { in canMoveAboveCall() 367 static bool canTransformAccumulatorRecursion(Instruction *I, CallInst *CI) { in canTransformAccumulatorRecursion() 432 CallInst *findTRECandidate(BasicBlock *BB, 435 void createTailRecurseLoopHeader(CallInst *CI); 439 bool eliminateCall(CallInst *CI); 452 CallInst *TailRecursionEliminator::findTRECandidate( in findTRECandidate() 461 CallInst *CI = nullptr; in findTRECandidate() [all …]
|
| H A D | LowerExpectIntrinsic.cpp | 58 getBranchWeight(Intrinsic::ID IntrinsicID, CallInst *CI, int BranchCount) { in getBranchWeight() 79 CallInst *CI = dyn_cast<CallInst>(SI.getCondition()); in handleSwitchExpect() 119 static void handlePhiDef(CallInst *Expect) { in handlePhiDef() 272 CallInst *CI; in handleBrSelExpect() 278 CI = dyn_cast<CallInst>(BSI.getCondition()); in handleBrSelExpect() 288 CI = dyn_cast<CallInst>(CmpI->getOperand(0)); in handleBrSelExpect() 362 CallInst *CI = dyn_cast<CallInst>(Inst); in lowerExpectIntrinsic()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/ |
| H A D | FastISel.h | 37 class CallInst; variable 493 bool lowerCallTo(const CallInst *CI, MCSymbol *Symbol, unsigned NumArgs); 494 bool lowerCallTo(const CallInst *CI, const char *SymName, 498 bool lowerCall(const CallInst *I); 504 bool selectStackmap(const CallInst *I); 505 bool selectPatchpoint(const CallInst *I); 512 bool selectXRayCustomEvent(const CallInst *II); 513 bool selectXRayTypedEvent(const CallInst *II); 553 const CallInst *CI, unsigned StartIdx); 554 bool lowerCallOperands(const CallInst *CI, unsigned ArgIdx, unsigned NumArgs,
|
| H A D | IntrinsicLowering.h | 21 class CallInst; variable 40 void LowerIntrinsicCall(CallInst *CI); 44 static bool LowerToByteSwap(CallInst *CI);
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Analysis/ |
| H A D | TypeMetadataUtils.cpp | 28 const CallInst *CI, DominatorTree &DT) { in findCallsAtConstantOffset() 42 } else if (auto *CI = dyn_cast<CallInst>(User)) { in findCallsAtConstantOffset() 55 int64_t Offset, const CallInst *CI, DominatorTree &DT) { in findLoadCallsAtConstantOffset() 77 SmallVectorImpl<CallInst *> &Assumes, const CallInst *CI, in findDevirtualizableCallsForTypeTest() 99 const CallInst *CI, DominatorTree &DT) { in findDevirtualizableCallsForTypeCheckedLoad()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/PowerPC/ |
| H A D | PPCBoolRetToInt.cpp | 82 if (CurrUser && !isa<CallInst>(Curr) && !isa<Constant>(Curr)) in findAllDefs() 143 return isa<ReturnInst>(V) || isa<CallInst>(V) || isa<PHINode>(V) || in getPromotablePHINodes() 147 return isa<Constant>(V) || isa<Argument>(V) || isa<CallInst>(V) || in getPromotablePHINodes() 210 if (auto *CI = dyn_cast<CallInst>(&I)) in runOnFunction() 233 !isa<Argument>(V) && !isa<CallInst>(V)) in runOnUse() 243 if (isa<CallInst>(U.getUser())) in runOnUse() 259 if (First && !isa<CallInst>(First) && !isa<Constant>(First)) in runOnUse()
|
| H A D | PPCLowerMASSVEntries.cpp | 58 bool handlePowSpecialCases(CallInst *CI, Function &Func, Module &M); 59 bool lowerMASSVCall(CallInst *CI, Function &Func, Module &M, 102 bool PPCLowerMASSVEntries::handlePowSpecialCases(CallInst *CI, Function &Func, in handlePowSpecialCases() 131 bool PPCLowerMASSVEntries::lowerMASSVCall(CallInst *CI, Function &Func, in lowerMASSVCall() 173 auto *CI = dyn_cast<CallInst>(User); in runOnModule()
|