| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/ |
| H A D | VNCoercion.cpp | 17 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, in canCoerceMustAliasedValueToLoad() argument 21 if (StoredTy == LoadTy) in canCoerceMustAliasedValueToLoad() 26 if (isFirstClassAggregateOrScalableType(LoadTy) || in canCoerceMustAliasedValueToLoad() 37 if (StoreSize < DL.getTypeSizeInBits(LoadTy).getFixedSize()) in canCoerceMustAliasedValueToLoad() 41 bool LoadNI = DL.isNonIntegralPointerType(LoadTy->getScalarType()); in canCoerceMustAliasedValueToLoad() 52 LoadTy->getPointerAddressSpace()) { in canCoerceMustAliasedValueToLoad() 60 if (StoredNI && StoreSize != DL.getTypeSizeInBits(LoadTy).getFixedSize()) in canCoerceMustAliasedValueToLoad() 176 static int analyzeLoadFromClobberingWrite(Type *LoadTy, Value *LoadPtr, in analyzeLoadFromClobberingWrite() argument 182 if (isFirstClassAggregateOrScalableType(LoadTy)) in analyzeLoadFromClobberingWrite() 192 uint64_t LoadSize = DL.getTypeSizeInBits(LoadTy).getFixedSize(); in analyzeLoadFromClobberingWrite() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Utils/ |
| H A D | VNCoercion.h | 37 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, 54 int analyzeLoadFromClobberingStore(Type *LoadTy, Value *LoadPtr, 62 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI, 70 int analyzeLoadFromClobberingMemInst(Type *LoadTy, Value *LoadPtr, 76 Value *getStoreValueForLoad(Value *SrcVal, unsigned Offset, Type *LoadTy, 81 Type *LoadTy, const DataLayout &DL); 87 Value *getLoadValueForLoad(LoadInst *SrcVal, unsigned Offset, Type *LoadTy, 92 Type *LoadTy, const DataLayout &DL); 99 Type *LoadTy, Instruction *InsertPt, 104 Type *LoadTy, const DataLayout &DL);
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Vectorize/ |
| H A D | LoadStoreVectorizer.cpp | 1148 Type *LoadTy = nullptr; in vectorizeLoadChain() local 1150 LoadTy = cast<LoadInst>(V)->getType(); in vectorizeLoadChain() 1151 if (LoadTy->isIntOrIntVectorTy()) in vectorizeLoadChain() 1154 if (LoadTy->isPtrOrPtrVectorTy()) { in vectorizeLoadChain() 1155 LoadTy = Type::getIntNTy(F.getParent()->getContext(), in vectorizeLoadChain() 1156 DL.getTypeSizeInBits(LoadTy)); in vectorizeLoadChain() 1160 assert(LoadTy && "Can't determine LoadInst type from chain"); in vectorizeLoadChain() 1162 unsigned Sz = DL.getTypeSizeInBits(LoadTy); in vectorizeLoadChain() 1195 auto *VecLoadTy = dyn_cast<FixedVectorType>(LoadTy); in vectorizeLoadChain() 1197 VecTy = FixedVectorType::get(LoadTy->getScalarType(), in vectorizeLoadChain() [all …]
|
| H A D | VectorCombine.cpp | 190 Type *LoadTy = Load->getType(); in vectorizeLoadInsert() local 192 TTI.getMemoryOpCost(Instruction::Load, LoadTy, Alignment, AS); in vectorizeLoadInsert()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Analysis/ |
| H A D | ConstantFolding.cpp | 548 Constant *FoldReinterpretLoadFromConstPtr(Constant *C, Type *LoadTy, in FoldReinterpretLoadFromConstPtr() argument 551 if (isa<ScalableVectorType>(LoadTy)) in FoldReinterpretLoadFromConstPtr() 555 auto *IntType = dyn_cast<IntegerType>(LoadTy); in FoldReinterpretLoadFromConstPtr() 566 if (LoadTy->isHalfTy()) in FoldReinterpretLoadFromConstPtr() 568 else if (LoadTy->isFloatTy()) in FoldReinterpretLoadFromConstPtr() 570 else if (LoadTy->isDoubleTy()) in FoldReinterpretLoadFromConstPtr() 572 else if (LoadTy->isVectorTy()) { in FoldReinterpretLoadFromConstPtr() 574 C->getContext(), DL.getTypeSizeInBits(LoadTy).getFixedSize()); in FoldReinterpretLoadFromConstPtr() 580 if (Res->isNullValue() && !LoadTy->isX86_MMXTy() && in FoldReinterpretLoadFromConstPtr() 581 !LoadTy->isX86_AMXTy()) in FoldReinterpretLoadFromConstPtr() [all …]
|
| H A D | Loads.cpp | 450 Type *LoadTy, in areNonOverlapSameBaseLoadAndStore() argument 462 auto LoadAccessSize = LocationSize::precise(DL.getTypeStoreSize(LoadTy)); in areNonOverlapSameBaseLoadAndStore()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/ARM/ |
| H A D | ARMParallelDSP.cpp | 229 LoadInst* CreateWideLoad(MemInstList &Loads, IntegerType *LoadTy); 737 IntegerType *LoadTy) { in CreateWideLoad() argument 777 LoadTy->getPointerTo(AddrSpace)); in CreateWideLoad() 778 LoadInst *WideLoad = IRB.CreateAlignedLoad(LoadTy, VecPtr, Base->getAlign()); in CreateWideLoad() 792 Value *ShiftVal = ConstantInt::get(LoadTy, OffsetTy->getBitWidth()); in CreateWideLoad()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/ |
| H A D | ScalarizeMaskedMemIntrin.cpp | 961 Type *LoadTy = CI->getType(); in optimizeCallInst() local 963 DL.getValueOrABITypeAlignment(MaybeAlign(AlignmentInt), LoadTy); in optimizeCallInst() 964 if (TTI.isLegalMaskedGather(LoadTy, Alignment)) in optimizeCallInst()
|
| H A D | GVN.cpp | 884 Type *LoadTy = Load->getType(); in MaterializeAdjustedValue() local 888 if (Res->getType() != LoadTy) { in MaterializeAdjustedValue() 889 Res = getStoreValueForLoad(Res, Offset, LoadTy, InsertPt, DL); in MaterializeAdjustedValue() 898 if (Load->getType() == LoadTy && Offset == 0) { in MaterializeAdjustedValue() 901 Res = getLoadValueForLoad(Load, Offset, LoadTy, InsertPt, DL); in MaterializeAdjustedValue() 914 Res = getMemInstValueForLoad(getMemIntrinValue(), Offset, LoadTy, in MaterializeAdjustedValue()
|
| H A D | SROA.cpp | 1270 Type *LoadTy = SomeLoad->getType(); in speculatePHINodeLoads() local 1272 PHINode *NewPN = PHIBuilder.CreatePHI(LoadTy, PN.getNumIncomingValues(), in speculatePHINodeLoads() 1307 LoadTy, InVal, Alignment, in speculatePHINodeLoads()
|
| /netbsd-src/external/apache2/llvm/dist/clang/include/clang/StaticAnalyzer/Core/PathSensitive/ |
| H A D | ExprEngine.h | 690 QualType LoadTy = QualType()); 762 QualType LoadTy);
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPURegisterBankInfo.cpp | 1143 const LLT LoadTy = MRI.getType(DstReg); in applyMappingLoad() local 1144 unsigned LoadSize = LoadTy.getSizeInBits(); in applyMappingLoad() 1162 (MemSize == 32 || LoadTy.isVector() || !isScalarLoadLegal(MI))) in applyMappingLoad() 1190 std::tie(Part64, Part32) = splitUnequalType(LoadTy, 64); in applyMappingLoad() 1194 auto Undef = B.buildUndef(LoadTy); in applyMappingLoad() 1195 auto Ins0 = B.buildInsert(LoadTy, Undef, Load0, 0); in applyMappingLoad() 1198 LLT WiderTy = widen96To128(LoadTy); in applyMappingLoad() 1226 unsigned NumSplitParts = LoadTy.getSizeInBits() / MaxNonSmrdLoadSize; in applyMappingLoad() 1227 const LLT LoadSplitTy = LoadTy.divide(NumSplitParts); in applyMappingLoad() 1232 if (LoadTy.isVector()) { in applyMappingLoad()
|
| H A D | AMDGPUISelLowering.cpp | 759 bool AMDGPUTargetLowering::isLoadBitCastBeneficial(EVT LoadTy, EVT CastTy, in isLoadBitCastBeneficial() argument 763 assert(LoadTy.getSizeInBits() == CastTy.getSizeInBits()); in isLoadBitCastBeneficial() 765 if (LoadTy.getScalarType() == MVT::i32) in isLoadBitCastBeneficial() 768 unsigned LScalarSize = LoadTy.getScalarSizeInBits(); in isLoadBitCastBeneficial()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombineLoadStoreAlloca.cpp | 538 static bool isMinMaxWithLoads(Value *V, Type *&LoadTy) { in isMinMaxWithLoads() argument 552 LoadTy = L1->getType(); in isMinMaxWithLoads()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/GlobalISel/ |
| H A D | MachineIRBuilder.cpp | 374 LLT LoadTy = Dst.getLLTTy(*getMRI()); in buildLoadFromOffset() local 376 getMF().getMachineMemOperand(&BaseMMO, Offset, LoadTy.getSizeInBytes()); in buildLoadFromOffset()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/Hexagon/ |
| H A D | HexagonISelLowering.cpp | 2949 MVT LoadTy = ty(Op); in LowerUnalignedLoad() local 2950 unsigned NeedAlign = Subtarget.getTypeAlignment(LoadTy); in LowerUnalignedLoad() 2989 assert(LoadTy.getSizeInBits() == 8*NeedAlign); in LowerUnalignedLoad() 3022 SDValue Load0 = DAG.getLoad(LoadTy, dl, Chain, Base0, WideMMO); in LowerUnalignedLoad() 3023 SDValue Load1 = DAG.getLoad(LoadTy, dl, Chain, Base1, WideMMO); in LowerUnalignedLoad() 3025 SDValue Aligned = DAG.getNode(HexagonISD::VALIGN, dl, LoadTy, in LowerUnalignedLoad()
|
| H A D | HexagonISelLoweringHVX.cpp | 1902 MVT LoadTy = MVT::getVectorVT(MVT::i8, HwLen); in WidenHvxLoad() local 1906 SDValue Load = DAG.getMaskedLoad(LoadTy, dl, Chain, Base, Offset, Mask, in WidenHvxLoad() 1907 DAG.getUNDEF(LoadTy), LoadTy, MemOp, in WidenHvxLoad()
|
| /netbsd-src/external/apache2/llvm/dist/clang/lib/StaticAnalyzer/Core/ |
| H A D | ExprEngine.cpp | 2964 QualType LoadTy) { in evalLoad() argument 2985 if (LoadTy.isNull()) in evalLoad() 2986 LoadTy = BoundEx->getType(); in evalLoad() 2987 V = state->getSVal(location.castAs<Loc>(), LoadTy); in evalLoad()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/SelectionDAG/ |
| H A D | LegalizeDAG.cpp | 1502 MVT LoadTy = TLI.getRegisterType(*DAG.getContext(), MVT::i8); in getSignAsIntValue() local 1504 SDValue StackPtr = DAG.CreateStackTemporary(FloatVT, LoadTy); in getSignAsIntValue() 1529 State.IntValue = DAG.getExtLoad(ISD::EXTLOAD, DL, LoadTy, State.Chain, IntPtr, in getSignAsIntValue() 1531 State.SignMask = APInt::getOneBitSet(LoadTy.getScalarSizeInBits(), 7); in getSignAsIntValue()
|
| H A D | SelectionDAGBuilder.cpp | 7517 Type *LoadTy = in getMemCmpLoad() local 7520 LoadTy = FixedVectorType::get(LoadTy, LoadVT.getVectorNumElements()); in getMemCmpLoad() 7523 PointerType::getUnqual(LoadTy)); in getMemCmpLoad() 7526 const_cast<Constant *>(LoadInput), LoadTy, *Builder.DL)) in getMemCmpLoad()
|