/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/MCTargetDesc/ |
H A D | X86ShuffleDecode.cpp | 48 void DecodeInsertElementMask(unsigned NumElts, unsigned Idx, unsigned Len, in DecodeInsertElementMask() argument 50 assert((Idx + Len) <= NumElts && "Insertion out of range"); in DecodeInsertElementMask() 52 for (unsigned i = 0; i != NumElts; ++i) in DecodeInsertElementMask() 55 ShuffleMask[Idx + i] = NumElts + i; in DecodeInsertElementMask() 76 void DecodeMOVSLDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask) { in DecodeMOVSLDUPMask() argument 77 for (int i = 0, e = NumElts / 2; i < e; ++i) { in DecodeMOVSLDUPMask() 83 void DecodeMOVSHDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask) { in DecodeMOVSHDUPMask() argument 84 for (int i = 0, e = NumElts / 2; i < e; ++i) { in DecodeMOVSHDUPMask() 90 void DecodeMOVDDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask) { in DecodeMOVDDUPMask() argument 93 for (unsigned l = 0; l < NumElts; l += NumLaneElts) in DecodeMOVDDUPMask() [all …]
|
H A D | X86ShuffleDecode.h | 35 void DecodeInsertElementMask(unsigned NumElts, unsigned Idx, unsigned Len, 46 void DecodeMOVSLDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask); 48 void DecodeMOVSHDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask); 50 void DecodeMOVDDUPMask(unsigned NumElts, SmallVectorImpl<int> &ShuffleMask); 52 void DecodePSLLDQMask(unsigned NumElts, unsigned Imm, 55 void DecodePSRLDQMask(unsigned NumElts, unsigned Imm, 58 void DecodePALIGNRMask(unsigned NumElts, unsigned Imm, 61 void DecodeVALIGNMask(unsigned NumElts, unsigned Imm, 65 void DecodePSHUFMask(unsigned NumElts, unsigned ScalarBits, unsigned Imm, 69 void DecodePSHUFHWMask(unsigned NumElts, unsigned Imm, [all …]
|
H A D | X86IntelInstPrinter.cpp | 162 unsigned NumElts; in printVecCompareInstr() local 164 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 8 : 16; in printVecCompareInstr() 166 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 4 : 8; in printVecCompareInstr() 168 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 2 : 4; in printVecCompareInstr() 169 OS << "{1to" << NumElts << "}"; in printVecCompareInstr() 302 unsigned NumElts; in printVecCompareInstr() local 304 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 8 : 16; in printVecCompareInstr() 306 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 4 : 8; in printVecCompareInstr() 308 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 2 : 4; in printVecCompareInstr() 309 OS << "{1to" << NumElts << "}"; in printVecCompareInstr()
|
H A D | X86ATTInstPrinter.cpp | 172 unsigned NumElts; in printVecCompareInstr() local 174 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 8 : 16; in printVecCompareInstr() 176 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 4 : 8; in printVecCompareInstr() 178 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 2 : 4; in printVecCompareInstr() 179 OS << "{1to" << NumElts << "}"; in printVecCompareInstr() 314 unsigned NumElts; in printVecCompareInstr() local 316 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 8 : 16; in printVecCompareInstr() 318 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 4 : 8; in printVecCompareInstr() 320 NumElts = (Desc.TSFlags & X86II::VEX_W) ? 2 : 4; in printVecCompareInstr() 321 OS << "{1to" << NumElts << "}"; in printVecCompareInstr()
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/ |
H A D | X86ShuffleDecodeConstantPool.cpp | 128 unsigned NumElts = Width / 8; in DecodePSHUFBMask() local 129 assert((NumElts == 16 || NumElts == 32 || NumElts == 64) && in DecodePSHUFBMask() 132 for (unsigned i = 0; i != NumElts; ++i) { in DecodePSHUFBMask() 167 unsigned NumElts = Width / ElSize; in DecodeVPERMILPMask() local 169 assert((NumElts == 2 || NumElts == 4 || NumElts == 8 || NumElts == 16) && in DecodeVPERMILPMask() 172 for (unsigned i = 0; i != NumElts; ++i) { in DecodeVPERMILPMask() 203 unsigned NumElts = Width / ElSize; in DecodeVPERMIL2PMask() local 205 assert((NumElts == 2 || NumElts == 4 || NumElts == 8) && in DecodeVPERMIL2PMask() 208 for (unsigned i = 0; i != NumElts; ++i) { in DecodeVPERMIL2PMask() 239 Index += Src * NumElts; in DecodeVPERMIL2PMask() [all …]
|
H A D | X86PartialReduction.cpp | 140 unsigned NumElts = MulTy->getNumElements(); in tryMAddReplacement() local 145 SmallVector<int, 16> EvenMask(NumElts / 2); in tryMAddReplacement() 146 SmallVector<int, 16> OddMask(NumElts / 2); in tryMAddReplacement() 147 for (int i = 0, e = NumElts / 2; i != e; ++i) { in tryMAddReplacement() 159 SmallVector<int, 32> ConcatMask(NumElts); in tryMAddReplacement() 215 unsigned NumElts = OpTy->getNumElements(); in trySADReplacement() local 219 if (ST->hasBWI() && NumElts >= 64) { in trySADReplacement() 222 } else if (ST->hasAVX2() && NumElts >= 32) { in trySADReplacement() 232 if (NumElts < 16) { in trySADReplacement() 235 for (unsigned i = 0; i != NumElts; ++i) in trySADReplacement() [all …]
|
H A D | X86MCInstLower.cpp | 2317 int NumElts; in addConstantComments() local 2320 case X86::MOVDDUPrm: NumElts = 2; break; in addConstantComments() 2321 case X86::VMOVDDUPrm: NumElts = 2; break; in addConstantComments() 2322 case X86::VMOVDDUPZ128rm: NumElts = 2; break; in addConstantComments() 2323 case X86::VBROADCASTSSrm: NumElts = 4; break; in addConstantComments() 2324 case X86::VBROADCASTSSYrm: NumElts = 8; break; in addConstantComments() 2325 case X86::VBROADCASTSSZ128rm: NumElts = 4; break; in addConstantComments() 2326 case X86::VBROADCASTSSZ256rm: NumElts = 8; break; in addConstantComments() 2327 case X86::VBROADCASTSSZrm: NumElts = 16; break; in addConstantComments() 2328 case X86::VBROADCASTSDYrm: NumElts = 4; break; in addConstantComments() [all …]
|
H A D | X86ISelLowering.cpp | 2117 handleMaskRegisterForCallingConv(unsigned NumElts, CallingConv::ID CC, in handleMaskRegisterForCallingConv() argument 2121 if (NumElts == 2) in handleMaskRegisterForCallingConv() 2123 if (NumElts == 4) in handleMaskRegisterForCallingConv() 2125 if (NumElts == 8 && CC != CallingConv::X86_RegCall && in handleMaskRegisterForCallingConv() 2128 if (NumElts == 16 && CC != CallingConv::X86_RegCall && in handleMaskRegisterForCallingConv() 2133 if (NumElts == 32 && (!Subtarget.hasBWI() || CC != CallingConv::X86_RegCall)) in handleMaskRegisterForCallingConv() 2136 if (NumElts == 64 && Subtarget.hasBWI() && CC != CallingConv::X86_RegCall) { in handleMaskRegisterForCallingConv() 2143 if (!isPowerOf2_32(NumElts) || (NumElts == 64 && !Subtarget.hasBWI()) || in handleMaskRegisterForCallingConv() 2144 NumElts > 64) in handleMaskRegisterForCallingConv() 2145 return {MVT::i8, NumElts}; in handleMaskRegisterForCallingConv() [all …]
|
H A D | X86InstCombineIntrinsic.cpp | 352 int NumElts = VT->getNumElements(); in simplifyX86varShift() local 375 for (int I = 0; I < NumElts; ++I) { in simplifyX86varShift() 523 unsigned NumElts = ArgTy->getNumElements(); in simplifyX86movmsk() local 525 Type *IntegerTy = Builder.getIntNTy(NumElts); in simplifyX86movmsk() 811 unsigned NumElts = VecTy->getNumElements(); in simplifyX86pshufb() local 812 assert((NumElts == 16 || NumElts == 32 || NumElts == 64) && in simplifyX86pshufb() 820 for (unsigned I = 0; I < NumElts; ++I) { in simplifyX86pshufb() 839 Index = ((Index < 0) ? NumElts : Index & 0x0F) + (I & 0xF0); in simplifyX86pshufb() 845 return Builder.CreateShuffleVector(V1, V2, makeArrayRef(Indexes, NumElts)); in simplifyX86pshufb() 856 unsigned NumElts = VecTy->getNumElements(); in simplifyX86vpermilvar() local [all …]
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Bitstream/Reader/ |
H A D | BitstreamReader.cpp | 101 unsigned NumElts = MaybeVBR.get(); in skipRecord() local 102 for (unsigned i = 0; i != NumElts; ++i) in skipRecord() 145 unsigned NumElts = MaybeNum.get(); in skipRecord() local 159 JumpToBit(GetCurrentBitNo() + static_cast<uint64_t>(NumElts) * in skipRecord() 165 for (; NumElts; --NumElts) in skipRecord() 173 if (Error Err = JumpToBit(GetCurrentBitNo() + NumElts * 6)) in skipRecord() 185 unsigned NumElts = MaybeNum.get(); in skipRecord() local 189 const size_t NewEnd = GetCurrentBitNo() + alignTo(NumElts, 4) * 8; in skipRecord() 216 uint32_t NumElts = MaybeNumElts.get(); in readRecord() local 217 Vals.reserve(Vals.size() + NumElts); in readRecord() [all …]
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/GISel/ |
H A D | AArch64PostLegalizerLowering.cpp | 66 static bool isREVMask(ArrayRef<int> M, unsigned EltSize, unsigned NumElts, in isREVMask() argument 81 for (unsigned i = 0; i < NumElts; ++i) { in isREVMask() 95 static bool isTRNMask(ArrayRef<int> M, unsigned NumElts, in isTRNMask() argument 97 if (NumElts % 2 != 0) in isTRNMask() 100 for (unsigned i = 0; i < NumElts; i += 2) { in isTRNMask() 103 static_cast<unsigned>(M[i + 1]) != i + NumElts + WhichResult)) in isTRNMask() 112 unsigned NumElts) { in getExtMask() argument 119 unsigned MaskBits = APInt(32, NumElts * 2).logBase2(); in getExtMask() 144 if (Imm < NumElts) in getExtMask() 147 Imm -= NumElts; in getExtMask() [all …]
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineVectorOps.cpp | 189 ElementCount NumElts = in foldBitcastExtElt() local 191 if (NumSrcElts == NumElts) in foldBitcastExtElt() 195 assert(NumSrcElts.isScalable() == NumElts.isScalable() && in foldBitcastExtElt() 200 if (NumSrcElts.getKnownMinValue() < NumElts.getKnownMinValue()) { in foldBitcastExtElt() 212 NumElts.getKnownMinValue() / NumSrcElts.getKnownMinValue(); in foldBitcastExtElt() 348 unsigned NumElts = EC.getKnownMinValue(); in visitExtractElementInst() local 352 if (!EC.isScalable() && IndexC->getValue().uge(NumElts)) in visitExtractElementInst() 358 if (!EC.isScalable() && NumElts != 1) { in visitExtractElementInst() 362 APInt UndefElts(NumElts, 0); in visitExtractElementInst() 363 APInt DemandedElts(NumElts, 0); in visitExtractElementInst() [all …]
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/IR/ |
H A D | AutoUpgrade.cpp | 1043 unsigned NumElts = ResultTy->getNumElements() * 8; in UpgradeX86PSLLDQIntrinsics() local 1046 Type *VecTy = FixedVectorType::get(Builder.getInt8Ty(), NumElts); in UpgradeX86PSLLDQIntrinsics() 1057 for (unsigned l = 0; l != NumElts; l += 16) in UpgradeX86PSLLDQIntrinsics() 1059 unsigned Idx = NumElts + i - Shift; in UpgradeX86PSLLDQIntrinsics() 1060 if (Idx < NumElts) in UpgradeX86PSLLDQIntrinsics() 1061 Idx -= NumElts - 16; // end of lane, switch operand. in UpgradeX86PSLLDQIntrinsics() 1065 Res = Builder.CreateShuffleVector(Res, Op, makeArrayRef(Idxs, NumElts)); in UpgradeX86PSLLDQIntrinsics() 1077 unsigned NumElts = ResultTy->getNumElements() * 8; in UpgradeX86PSRLDQIntrinsics() local 1080 Type *VecTy = FixedVectorType::get(Builder.getInt8Ty(), NumElts); in UpgradeX86PSRLDQIntrinsics() 1091 for (unsigned l = 0; l != NumElts; l += 16) in UpgradeX86PSRLDQIntrinsics() [all …]
|
H A D | IRBuilder.cpp | 546 ElementCount NumElts = PtrsTy->getElementCount(); in CreateMaskedGather() local 547 auto *DataTy = VectorType::get(PtrTy->getElementType(), NumElts); in CreateMaskedGather() 551 VectorType::get(Type::getInt1Ty(Context), NumElts)); in CreateMaskedGather() 576 ElementCount NumElts = PtrsTy->getElementCount(); in CreateMaskedScatter() local 580 assert(NumElts == DataTy->getElementCount() && in CreateMaskedScatter() 587 VectorType::get(Type::getInt1Ty(Context), NumElts)); in CreateMaskedScatter() 1024 int NumElts = Ty->getElementCount().getKnownMinValue(); in CreateVectorReverse() local 1025 for (int i = 0; i < NumElts; ++i) in CreateVectorReverse() 1026 ShuffleMask.push_back(NumElts - i - 1); in CreateVectorReverse() 1045 unsigned NumElts = cast<FixedVectorType>(V1->getType())->getNumElements(); in CreateVectorSplice() local [all …]
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/SelectionDAG/ |
H A D | LegalizeTypesGeneric.cpp | 347 unsigned NumElts = 2; in ExpandOp_BITCAST() local 351 NumElts); in ExpandOp_BITCAST() 355 NumElts = N->getValueType(0).getVectorNumElements(); in ExpandOp_BITCAST() 360 IntegerToVector(N->getOperand(0), NumElts, Ops, NVT.getVectorElementType()); in ExpandOp_BITCAST() 363 DAG.getBuildVector(NVT, dl, makeArrayRef(Ops.data(), NumElts)); in ExpandOp_BITCAST() 374 unsigned NumElts = VecVT.getVectorNumElements(); in ExpandOp_BUILD_VECTOR() local 385 NewElts.reserve(NumElts*2); in ExpandOp_BUILD_VECTOR() 387 for (unsigned i = 0; i < NumElts; ++i) { in ExpandOp_BUILD_VECTOR() 412 unsigned NumElts = VecVT.getVectorNumElements(); in ExpandOp_INSERT_VECTOR_ELT() local 424 EVT NewVecVT = EVT::getVectorVT(*DAG.getContext(), NewEVT, NumElts*2); in ExpandOp_INSERT_VECTOR_ELT() [all …]
|
H A D | LegalizeVectorTypes.cpp | 3281 unsigned NumElts = VT.getVectorNumElements(); in WidenVecRes_BinaryCanTrap() local 3283 while (!TLI.isTypeLegal(VT) && NumElts != 1) { in WidenVecRes_BinaryCanTrap() 3284 NumElts = NumElts / 2; in WidenVecRes_BinaryCanTrap() 3285 VT = EVT::getVectorVT(*DAG.getContext(), WidenEltVT, NumElts); in WidenVecRes_BinaryCanTrap() 3288 if (NumElts != 1 && !TLI.canOpTrap(N->getOpcode(), VT)) { in WidenVecRes_BinaryCanTrap() 3296 if (NumElts == 1) in WidenVecRes_BinaryCanTrap() 3315 while (CurNumElts >= NumElts) { in WidenVecRes_BinaryCanTrap() 3321 Idx += NumElts; in WidenVecRes_BinaryCanTrap() 3322 CurNumElts -= NumElts; in WidenVecRes_BinaryCanTrap() 3325 NumElts = NumElts / 2; in WidenVecRes_BinaryCanTrap() [all …]
|
/netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/ADT/ |
H A D | SmallVector.h | 382 void growAndAssign(size_t NumElts, const T &Elt) { in growAndAssign() argument 385 T *NewElts = mallocForGrow(NumElts, NewCapacity); in growAndAssign() 386 std::uninitialized_fill_n(NewElts, NumElts, Elt); in growAndAssign() 389 this->set_size(NumElts); in growAndAssign() 528 void growAndAssign(size_t NumElts, T Elt) { in growAndAssign() argument 532 this->grow(NumElts); in growAndAssign() 533 std::uninitialized_fill_n(this->begin(), NumElts, Elt); in growAndAssign() 534 this->set_size(NumElts); in growAndAssign() 669 void assign(size_type NumElts, ValueParamT Elt) { in assign() argument 671 if (NumElts > this->capacity()) { in assign() [all …]
|
/netbsd-src/external/apache2/llvm/dist/clang/include/clang/AST/ |
H A D | APValue.h | 271 unsigned NumElts; 272 Vec() : Elts(nullptr), NumElts(0) {} 277 unsigned NumElts, ArrSize; 278 Arr(unsigned NumElts, unsigned ArrSize); 499 return ((const Vec *)(const void *)&Data)->NumElts; 523 return ((const Arr *)(const void *)&Data)->NumElts; 683 V->NumElts = N; 684 return {V->Elts, V->NumElts};
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/ARM/ |
H A D | MVELaneInterleavingPass.cpp | 288 unsigned NumElts = VT->getNumElements(); in tryInterleave() local 292 if (BaseElts == 0 || NumElts % BaseElts != 0) { in tryInterleave() 323 for (unsigned Base = 0; Base < NumElts; Base += BaseElts) { in tryInterleave() 329 for (unsigned Base = 0; Base < NumElts; Base += BaseElts) { in tryInterleave()
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Instrumentation/ |
H A D | PoisonChecking.cpp | 205 unsigned NumElts = VecVTy->getNumElements(); in generateCreationChecks() local 208 ConstantInt::get(Idx->getType(), NumElts)); in generateCreationChecks() 218 unsigned NumElts = VecVTy->getNumElements(); in generateCreationChecks() local 221 ConstantInt::get(Idx->getType(), NumElts)); in generateCreationChecks()
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/ |
H A D | TargetLoweringObjectFile.cpp | 109 unsigned NumElts = CDS->getNumElements(); in IsNullTerminatedString() local 110 assert(NumElts != 0 && "Can't have an empty CDS"); in IsNullTerminatedString() 112 if (CDS->getElementAsInteger(NumElts-1) != 0) in IsNullTerminatedString() 116 for (unsigned i = 0; i != NumElts-1; ++i) in IsNullTerminatedString()
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/ |
H A D | RelLookupTableConverter.cpp | 95 unsigned NumElts = LookupTableArr->getType()->getNumElements(); in createRelLookupTable() local 97 ArrayType::get(Type::getInt32Ty(M.getContext()), NumElts); in createRelLookupTable() 106 SmallVector<Constant *, 64> RelLookupTableContents(NumElts); in createRelLookupTable()
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/NVPTX/ |
H A D | NVPTXISelLowering.cpp | 187 unsigned NumElts = VT.getVectorNumElements(); in ComputePTXValueVTs() local 192 if (EltVT == MVT::f16 && NumElts % 2 == 0) { in ComputePTXValueVTs() 194 NumElts /= 2; in ComputePTXValueVTs() 196 for (unsigned j = 0; j != NumElts; ++j) { in ComputePTXValueVTs() 237 unsigned NumElts = AccessSize / EltSize; in CanMergeParamLoadStoresStartingAt() local 239 if (AccessSize != EltSize * NumElts) in CanMergeParamLoadStoresStartingAt() 243 if (Idx + NumElts > ValueVTs.size()) in CanMergeParamLoadStoresStartingAt() 247 if (NumElts != 4 && NumElts != 2) in CanMergeParamLoadStoresStartingAt() 250 for (unsigned j = Idx + 1; j < Idx + NumElts; ++j) { in CanMergeParamLoadStoresStartingAt() 260 return NumElts; in CanMergeParamLoadStoresStartingAt() [all …]
|
/netbsd-src/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
H A D | InterleavedAccessPass.cpp | 207 unsigned NumElts = Mask.size(); in isReInterleaveMask() local 208 if (NumElts < 4) in isReInterleaveMask() 213 if (NumElts % Factor) in isReInterleaveMask() 216 unsigned LaneLen = NumElts / Factor; in isReInterleaveMask()
|
H A D | HardwareLoops.cpp | 179 PHINode *InsertPHICounter(Value *NumElts, Value *EltsRem); 500 PHINode* HardwareLoop::InsertPHICounter(Value *NumElts, Value *EltsRem) { in InsertPHICounter() argument 505 PHINode *Index = Builder.CreatePHI(NumElts->getType(), 2); in InsertPHICounter() 506 Index->addIncoming(NumElts, Preheader); in InsertPHICounter()
|