/freebsd-src/contrib/llvm-project/llvm/lib/Target/RISCV/ |
H A D | RISCVGatherScatterLowering.cpp | 44 // Cache of the BasePtr and Stride determined from this GEP. When a GEP is 73 bool matchStridedRecurrence(Value *Index, Loop *L, Value *&Stride, 89 // TODO: Should we consider the mask when looking for a stride? 117 Value *Stride = ConstantInt::get(StartVal->getType(), StrideVal); in matchStridedConstant() local 119 return std::make_pair(StartVal, Stride); in matchStridedConstant() 158 Value *Stride; in matchStridedStart() local 159 std::tie(Start, Stride) = matchStridedStart(BO->getOperand(OtherIndex), in matchStridedStart() 166 // Add the splat value to the start or multiply the start and stride by the in matchStridedStart() 180 Stride = Builder.CreateMul(Stride, Spla in matchStridedStart() 196 matchStridedRecurrence(Value * Index,Loop * L,Value * & Stride,PHINode * & BasePtr,BinaryOperator * & Inc,IRBuilderBase & Builder) matchStridedRecurrence() argument 491 Value *BasePtr, *Stride; tryCreateStridedLoadStore() local [all...] |
/freebsd-src/contrib/pnglite/ |
H A D | pnglite.c | 382 png_filter_sub(unsigned stride, uint8_t *in, uint8_t *out, unsigned len) in png_filter_sub() argument 388 if (i >= stride) in png_filter_sub() 389 a = out[i - stride]; in png_filter_sub() 396 png_filter_up(unsigned stride __unused, uint8_t *in, uint8_t *out, in png_filter_up() 409 png_filter_average(unsigned stride, uint8_t *in, uint8_t *out, in png_filter_average() argument 421 if (i >= stride) in png_filter_average() 422 a = out[i - stride]; in png_filter_average() 452 png_filter_paeth(unsigned stride, uint8_t *in, uint8_t *out, uint8_t *prev_line, in png_filter_paeth() argument 461 if (prev_line && i >= stride) { in png_filter_paeth() 462 a = out[i - stride]; in png_filter_paeth() [all …]
|
/freebsd-src/contrib/llvm-project/clang/lib/Headers/ |
H A D | amxintrin.h | 72 /// Load tile rows from memory specifieid by "base" address and "stride" into 84 /// \param stride 85 /// The stride between the rows' data to be loaded in memory. 86 #define _tile_loadd(dst, base, stride) \ argument 88 (__SIZE_TYPE__)(stride)) 90 /// Load tile rows from memory specifieid by "base" address and "stride" into 104 /// \param stride 105 /// The stride between the rows' data to be loaded in memory. 106 #define _tile_stream_loadd(dst, base, stride) \ argument 108 (__SIZE_TYPE__)(stride)) [all …]
|
/freebsd-src/sys/contrib/openzfs/tests/zfs-tests/cmd/ |
H A D | stride_dd.c | 29 static off_t stride = 1; 47 " [-s stride] [-k seekblocks] [-K seekbytes]\n" in usage() 51 "Simplified version of dd that supports the stride option.\n" in usage() 52 "A stride of n means that for each block written, n - 1 blocks\n" in usage() 53 "are skipped in both the input and output file. A stride of 1\n" in usage() 62 " stride: Read/write a block then skip (stride - 1) blocks" in parse_options() 145 stride = atoi(optarg); in main() 183 if (bsize <= 0 || stride <= 0 || ifile == NULL || ofile == NULL || in main() 215 if (stride > in main() 28 static off_t stride = 0; global() variable [all...] |
/freebsd-src/contrib/llvm-project/clang/include/clang/AST/ |
H A D | ExprOpenMP.h | 57 enum { BASE, LOWER_BOUND, LENGTH, STRIDE, END_EXPR }; global() enumerator 64 OMPArraySectionExpr(Expr * Base,Expr * LowerBound,Expr * Length,Expr * Stride,QualType Type,ExprValueKind VK,ExprObjectKind OK,SourceLocation ColonLocFirst,SourceLocation ColonLocSecond,SourceLocation RBracketLoc) OMPArraySectionExpr() argument
|
/freebsd-src/contrib/llvm-project/llvm/lib/Target/VE/ |
H A D | VVPInstrPatternsVec.td | 23 // Unmasked (imm stride). 26 (i64 simm7:$stride), (MaskVT true_mask), i32:$avl), 28 (LO7 $stride), $addr, $val, $avl)>; 32 i64:$stride, (MaskVT true_mask), i32:$avl), 34 $stride, $addr, $val, $avl)>; 35 // Masked (imm stride). 38 (i64 simm7:$stride), MaskVT:$mask, i32:$avl), 40 (LO7 $stride), $addr, $val, $mask, $avl)>; 44 i64:$stride, MaskVT:$mask, i32:$avl), 46 $stride, $addr, $val, $mask, $avl)>; [all …]
|
/freebsd-src/contrib/ofed/libmlx5/ |
H A D | mlx5dv_init_obj.3 | 29 uint32_t stride; 36 uint32_t stride; 65 uint32_t stride; 77 uint32_t stride;
|
/freebsd-src/contrib/llvm-project/llvm/include/llvm/Analysis/ |
H A D | VectorUtils.h | 381 /// Create a stride shuffle mask. 384 /// are incremented by \p Stride. The mask can be used to deinterleave an 388 /// <Start, Start + Stride, ..., Start + Stride * (VF - 1)> 390 /// For example, the mask for Start = 0, Stride = 2, and VF = 4 is: 393 llvm::SmallVector<int, 16> createStrideMask(unsigned Start, unsigned Stride, 444 /// The group of interleaved loads/stores sharing the same stride and 449 /// value of the access's stride. 476 InterleaveGroup(InstTy *Instr, int32_t Stride, Align Alignment) in insertMember() 478 Factor = std::abs(Stride); in insertMember() 450 InterleaveGroup(InstTy * Instr,int32_t Stride,Align Alignment) InterleaveGroup() argument 686 int64_t Stride = 0; global() member 706 createInterleaveGroup(Instruction * Instr,int Stride,Align Alignment) createInterleaveGroup() argument [all...] |
H A D | LoopCacheAnalysis.h | 91 /// - equal to '(TripCount * stride) / cache_line_size' if: 92 /// + the reference stride is less than the cache line size, and 111 /// the loop induction variable is the rightmost one, and the access stride is 112 /// smaller than the cache line size \p CLS. Provide a valid \p Stride value 114 bool isConsecutive(const Loop &L, const SCEV *&Stride, unsigned CLS) const; 184 /// reference stride if the reference stride is less than the cache line 187 /// - equal to the innermost loop trip count if the reference stride is greater 245 /// - equal to '(TripCount * stride) / cache_line_size' if (a) loop \p L's 247 /// with loop \p L, and (b) the reference stride is less than the cache
|
/freebsd-src/contrib/llvm-project/llvm/include/llvm/Object/ |
H A D | DXContainer.h | 48 uint32_t Stride = sizeof(T); // size of each element in the list. member 51 ViewArray(StringRef D, size_t S) : Data(D), Stride(S) {} in ViewArray() 60 uint32_t Stride; // size of each element in the list. member 64 : Data(A.Data), Stride(A.Stride), Current(C) {} in iterator() 75 memcpy(static_cast<void *>(&Val), Current, std::min(Stride, MaxStride())); 83 Current += Stride; 95 Current -= Stride; 113 size_t size() const { return Data.size() / Stride; } in size() 162 uint32_t getResourceStride() const { return Resources.Stride; } [all...] |
H A D | Minidump.h | 111 MemoryInfoIterator(ArrayRef<uint8_t> Storage, size_t Stride) in MemoryInfoIterator() argument 112 : Storage(Storage), Stride(Stride) { in MemoryInfoIterator() 113 assert(Storage.size() % Stride == 0); in MemoryInfoIterator() 126 Storage = Storage.drop_front(Stride); 132 size_t Stride; variable
|
/freebsd-src/contrib/llvm-project/libcxx/include/__mdspan/ |
H A D | layout_stride.h | 134 // compute the permutation for sorting the stride array 135 // we never actually sort the stride array 156 // [mdspan.layout.stride.cons], constructors 229 // stride() only compiles for rank > 0 in requires() 232 static_cast<index_type>(__other.stride(_Pos))...}; in requires() 237 // stride() only compiles for rank > 0 241 return ((static_cast<index_type>(__other.stride(_Pos)) > static_cast<index_type>(0)) && ... && true); 254 // [mdspan.layout.stride.obs], observers 330 // according to the standard layout_stride does not have a constraint on stride(r) for rank>0 332 _LIBCPP_HIDE_FROM_ABI constexpr index_type stride(rank_typ function [all...] |
/freebsd-src/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | LoopAccessAnalysis.cpp | 124 cl::desc("Enable symbolic stride memory access versioning")); 139 "laa-speculate-unit-stride", cl::Hidden, 160 // symbolic stride replaced by one. in replaceSymbolicStrideSCEV() 163 // For a non-symbolic stride, just return the original expression. in replaceSymbolicStrideSCEV() 840 int64_t Stride = getPtrStride(PSE, AccessTy, Ptr, L, Strides).value_or(0); in isNoWrap() 841 if (Stride == 1 || PSE.hasNoOverflow(Ptr, SCEVWrapPredicate::IncrementNUSW)) in isNoWrap() 1456 /// Check whether the access through \p Ptr has a constant stride. in isNoWrapAddRec() 1469 LLVM_DEBUG(dbgs() << "LAA: Bad stride - Scalable object: " << *AccessTy in getPtrStride() 1479 LLVM_DEBUG(dbgs() << "LAA: Bad stride - Not an AddRecExpr pointer " << *Ptr in getPtrStride() 1484 // The access function must stride ove in getPtrStride() 842 int64_t Stride = getPtrStride(PSE, AccessTy, Ptr, L, Strides).value_or(0); isNoWrap() local 1518 int64_t Stride = StepVal / Size; getPtrStride() local 1821 isSafeDependenceDistance(const DataLayout & DL,ScalarEvolution & SE,const SCEV & BackedgeTakenCount,const SCEV & Dist,uint64_t Stride,uint64_t TypeByteSize) isSafeDependenceDistance() argument 1879 areStridedAccessesIndependent(uint64_t Distance,uint64_t Stride,uint64_t TypeByteSize) areStridedAccessesIndependent() argument 1995 uint64_t Stride = std::abs(StrideAPtr); getDependenceDistanceStrideAndSize() local [all...] |
H A D | LoopCacheAnalysis.cpp | 291 const SCEV *Stride = nullptr; in computeRefCost() local 292 if (isConsecutive(L, Stride, CLS)) { in computeRefCost() 294 // (TripCount*Stride)/CLS. in computeRefCost() 295 assert(Stride != nullptr && in computeRefCost() 296 "Stride should not be null for consecutive access!"); in computeRefCost() 297 Type *WiderType = SE.getWiderType(Stride->getType(), TripCount->getType()); in computeRefCost() 299 Stride = SE.getNoopOrAnyExtend(Stride, WiderType); in computeRefCost() 301 const SCEV *Numerator = SE.getMulExpr(Stride, TripCount); in computeRefCost() 310 << "Access is consecutive: RefCost=(TripCount*Stride)/CL in computeRefCost() 462 isConsecutive(const Loop & L,const SCEV * & Stride,unsigned CLS) const isConsecutive() argument [all...] |
/freebsd-src/contrib/llvm-project/llvm/lib/Object/ |
H A D | DXContainer.cpp | 290 if (Error Err = readInteger(Data, Current, Resources.Stride)) in parse() 294 size_t BindingDataSize = Resources.Stride * ResourceCount; in parse() 303 Resources.Stride = sizeof(v2::ResourceBindInfo); in parse() 344 if (Error Err = readInteger(Data, Current, SigInputElements.Stride)) in parse() 347 // Assign the stride to all the arrays. in parse() 348 SigOutputElements.Stride = SigPatchOrPrimElements.Stride = in parse() 349 SigInputElements.Stride; in parse() 352 (ptrdiff_t)(ElementCount * SigInputElements.Stride)) in parse() 356 size_t InputSize = SigInputElements.Stride * InputCoun in parse() [all...] |
/freebsd-src/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | StraightLineStrengthReduce.cpp | 144 : CandidateKind(CT), Base(B), Index(Idx), Stride(S), Ins(I) {} in Candidate() 150 // Note that Index and Stride of a GEP candidate do not necessarily have the 151 // same integer type. In that case, during rewriting, Stride will be 155 Value *Stride = nullptr; member 166 // <Base: a, Index: 1, Stride: b + 2> 170 // <Base: b, Index: 2, Stride: a + 1> 182 // share the same base and stride. 230 // A helper function that factors ArrayIdx to a product of a stride and a 276 // They share the same base, stride, and candidate kind. in isBasisFor() 277 Basis.Base == C.Base && Basis.Stride in isBasisFor() 289 isAddFoldable(const SCEV * Base,ConstantInt * Index,Value * Stride,TargetTransformInfo * TTI) isAddFoldable() argument [all...] |
/freebsd-src/contrib/llvm-project/llvm/lib/Target/X86/ |
H A D | X86InterleavedAccess.cpp | 60 /// Reference to the interleaving stride in terms of elements. 106 /// \p 'Ind' and the interleaving stride factor \p F. In order to generate 135 // Stride 4: in isSupported() 138 // Stride 3: in isSupported() 150 // We support shuffle represents stride 4 for byte type with size of in isSupported() 198 // In the case of stride 3 with a vector of 32 elements load the information in decompose() 289 unsigned VecElems, unsigned Stride, in reorderSubVector() argument 293 for (unsigned i = 0; i < Stride; i++) in reorderSubVector() 301 for (unsigned i = 0; i < (VecElems / 16) * Stride; i += 2) { in reorderSubVector() 302 genShuffleBland(VT, VPShuf, OptimizeShuf, (i / Stride) * 1 in reorderSubVector() 436 createShuffleStride(MVT VT,int Stride,SmallVectorImpl<int> & Mask) createShuffleStride() argument [all...] |
H A D | X86LowerAMXType.cpp | 252 // Use the maximun column as stride. in combineLoadBitcast() 253 Value *Stride = Builder.getInt64(64); in combineLoadBitcast() 255 std::array<Value *, 4> Args = {Row, Col, I8Ptr, Stride}; 263 // %stride); in combineBitcastStore() 278 // Use the maximum column as stride. It must be the same with load in combineBitcastStore() 279 // stride. in combineBitcastStore() 280 Value *Stride = Builder.getInt64(64); in combineBitcastStore() 282 std::array<Value *, 5> Args = {Row, Col, I8Ptr, Stride, Tile}; in combineBitcastStore() 304 Value *I8Ptr, *Stride; in transformBitcast() 310 Stride in transformBitcast() 246 Value *Stride = Builder.getInt64(64); combineLoadBitcast() local 273 Value *Stride = Builder.getInt64(64); combineBitcastStore() local 297 Value *I8Ptr, *Stride; transformBitcast() local 474 Value *Stride = Builder.getInt64(64); createTileStore() local 499 Value *Stride = Builder.getInt64(64); replaceWithTileLoad() local 936 Value *Stride = Builder.CreateSExt(Col, Builder.getInt64Ty()); combineCastStore() local 962 Value *Stride = Builder.CreateSExt(Col, Builder.getInt64Ty()); combineLoadCast() local 1134 Value *I8Ptr, *Stride; transformAMXCast() local [all...] |
/freebsd-src/contrib/llvm-project/openmp/runtime/src/ |
H A D | kmp_gsupport.cpp | 580 // "*p_ub += stride" compensates for the discrepancy. 583 // stride value. We adjust the dispatch parameters accordingly (by +-1), but 584 // we still adjust p_ub by the actual stride value. 605 long stride; \ 624 (kmp_int *)p_ub, (kmp_int *)&stride); \ 627 KMP_DEBUG_ASSERT(stride == str); \ 645 long stride; \ 664 (kmp_int *)p_ub, (kmp_int *)&stride); \ 667 KMP_DEBUG_ASSERT(stride == str); \ 690 long stride; \ 1366 kmp_int lb, ub, stride; KMP_EXPAND_NAME() local 1389 kmp_int lb, ub, stride; KMP_EXPAND_NAME() local [all...] |
/freebsd-src/contrib/llvm-project/llvm/include/llvm/IR/ |
H A D | MatrixBuilder.h | 65 /// \p Stride - Space between columns 67 Value *Stride, bool IsVolatile, unsigned Rows, 71 Value *Ops[] = {DataPtr, Stride, B.getInt1(IsVolatile), B.getInt32(Rows), 73 Type *OverloadedTypes[] = {RetType, Stride->getType()}; 88 /// \p Stride - Space between columns 90 Value *Stride, bool IsVolatile, 94 Stride, B.getInt1(IsVolatile), 96 Type *OverloadedTypes[] = {Matrix->getType(), Stride->getType()};
|
/freebsd-src/sys/dev/vt/hw/simplefb/ |
H A D | simplefb.c | 164 pcell_t height, width, stride; in vt_simplefb_init() 178 OF_getencprop(node, "stride", &stride, sizeof(stride)) == -1 || in vt_simplefb_init() 185 sc->fb_stride = stride; in vt_simplefb_init() 162 pcell_t height, width, stride; vt_simplefb_init() local
|
/freebsd-src/crypto/openssh/openbsd-compat/ |
H A D | bcrypt_pbkdf.c | 122 size_t i, j, amt, stride; in bcrypt_pbkdf() local 134 stride = (keylen + sizeof(out) - 1) / sizeof(out); in bcrypt_pbkdf() 135 amt = (keylen + stride - 1) / stride; in bcrypt_pbkdf() 168 size_t dest = i * stride + (count - 1); in bcrypt_pbkdf()
|
/freebsd-src/stand/efi/loader/ |
H A D | framebuffer.c | 336 uint32_t horiz, vert, stride; in efifb_from_uga() 366 * the stride. Our only option is to look at the system and in efifb_from_uga() 381 stride = 1728; in efifb_from_uga() 389 stride = 2048; in efifb_from_uga() 401 efifb->fb_stride = stride; in efifb_from_uga() 410 * The stride is equal or larger to the width. Often it's the in efifb_from_uga() 426 "hw.efifb.stride.\n"); in efifb_from_uga() 463 ev = getenv("hw.efifb.stride"); in efifb_from_uga() 466 /* Determine the stride. */ in efifb_from_uga() 472 printf("Unable to reliably detect the stride in efifb_from_uga() 327 uint32_t horiz, vert, stride; efifb_from_uga() local [all...] |
/freebsd-src/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/ |
H A D | AggressiveInstCombine.cpp | 838 // Calculate GEP Stride and accumulated const ModOffset. Return Stride and 843 std::optional<APInt> Stride; in getStrideAndModOffsetOfGEP() 845 // Return a minimum gep stride, greatest common divisor of consective gep in getStrideAndModOffsetOfGEP() 857 if (!Stride) in getStrideAndModOffsetOfGEP() 858 Stride = Scale; in getStrideAndModOffsetOfGEP() 860 Stride = APIntOps::GreatestCommonDivisor(*Stride, Scale); in getStrideAndModOffsetOfGEP() 868 if (!isa<GlobalVariable>(PtrOp) || !Stride) in getStrideAndModOffsetOfGEP() 872 // remainder of division by minimum GEP stride in getStrideAndModOffsetOfGEP() 841 std::optional<APInt> Stride; getStrideAndModOffsetOfGEP() local [all...] |
/freebsd-src/contrib/llvm-project/libcxx/include/ |
H A D | mdspan | 133 constexpr index_type stride(rank_type) const noexcept; 187 constexpr index_type stride(rank_type) const noexcept; 213 // [mdspan.layout.stride.cons], constructors 226 // [mdspan.layout.stride.obs], observers 243 constexpr index_type stride(rank_type i) const noexcept { return strides_[i]; } 355 constexpr index_type stride(rank_type r) const 356 { return map_.stride(r); }
|