/llvm-project/mlir/lib/Dialect/NVGPU/Transforms/ |
H A D | OptimizeSharedMemory.cpp | 49 int64_t srcDim, int64_t tgtDim) { in permuteVectorOffset() argument 52 Value src = indices[srcDim]; in permuteVectorOffset() 104 MemRefType memrefTy, int64_t srcDim, in transformIndices() argument 107 permuteVectorOffset(builder, loc, indices, memrefTy, srcDim, tgtDim); in transformIndices() 191 int64_t srcDim = memRefType.getRank() - 2; in optimizeSharedMemoryReadsAndWrites() 202 memRefType, srcDim, tgtDim); in optimizeSharedMemoryReadsAndWrites() 215 memRefType, srcDim, tgtDim); in optimizeSharedMemoryReadsAndWrites() 187 int64_t srcDim = memRefType.getRank() - 2; optimizeSharedMemoryReadsAndWrites() local
|
/llvm-project/mlir/lib/Dialect/Affine/Utils/ |
H A D | ViewLikeInterfaceUtils.cpp | 119 for (int64_t srcDim = 0; srcDim < srcRank; ++srcDim) { in resolveSizesIntoOpWithSizes() local 120 if (rankReducedSourceDims[srcDim]) { in resolveSizesIntoOpWithSizes() 121 resolvedSizes.push_back(sourceSizes[srcDim]); in resolveSizesIntoOpWithSizes()
|
/llvm-project/mlir/lib/Dialect/Tensor/Utils/ |
H A D | Utils.cpp | 152 int64_t srcDim = 0; in isCastLikeInsertSliceOp() 165 {op.getSource(), srcDim}, {op.getResult(), resultDim}); in isCastLikeExtractSliceOp() 168 ++srcDim; in isCastLikeExtractSliceOp() 133 int64_t srcDim = 0; isCastLikeInsertSliceOp() local
|
/llvm-project/mlir/lib/Dialect/Tensor/Transforms/ |
H A D | FoldTensorSubsetOps.cpp | 199 int64_t srcDim = 0; in matchAndRewrite() 205 sourceInsertSliceOp.getMixedSizes()[srcDim++]) { in matchAndRewrite() 179 int64_t srcDim = 0; matchAndRewrite() local
|
H A D | BufferizableOpInterfaceImpl.cpp | 778 Value srcDim = rewriter.create<tensor::DimOp>(loc, padOp.getSource(), i); in bufferize() 785 loc, sumExpr, ValueRange{srcDim, lowPad, highPad}); in bufferize() 780 Value srcDim = rewriter.create<tensor::DimOp>(loc, padOp.getSource(), i); bufferize() local
|
/llvm-project/flang/runtime/ |
H A D | namelist.cpp | 385 const auto &srcDim{source.GetDimension(j)}; in HandleComponent() local 387 .SetBounds(1, srcDim.UpperBound()) in HandleComponent() 388 .SetByteStride(srcDim.ByteStride()); in HandleComponent()
|
/llvm-project/mlir/lib/Dialect/SparseTensor/Transforms/Utils/ |
H A D | CodegenUtils.cpp | 244 auto srcDim = srcShape[i]; in genReshapeDstShape() local 249 // For example, if srcDim = 8, then the expanded shape could be <2x?x2>, in genReshapeDstShape() 253 // by dividing srcDim by the product of the static dimensions. in genReshapeDstShape() 263 builder.create<arith::DivUIOp>(loc, srcDim, productVal); in genReshapeDstShape()
|
/llvm-project/mlir/lib/Dialect/Vector/IR/ |
H A D | VectorOps.cpp | 2507 int64_t srcDim = srcVectorType.getDimSize(dimIdx); in verify() 2509 if (srcDim != 1 && srcDim != dstDim) in verify() 2515 if ((srcDim == 1 && srcDimScalableFlag && dstDim != 1) || in inferReturnTypes() 2519 (srcDim != 1 || srcDimScalableFlag))) in inferReturnTypes() 2524 mismatchingDims->first.dim = srcDim; in inferReturnTypes() 2395 int64_t srcDim = srcVectorType.getDimSize(r); isBroadcastableTo() local
|
/llvm-project/mlir/lib/Dialect/Tensor/IR/ |
H A D | TensorOps.cpp | 1984 int64_t srcDim = expandShapeOp.getCorrespondingSourceDim(*dim); in getCanonicalizationPatterns() 1989 ReassociationIndices grp = expandShapeOp.getReassociationIndices()[srcDim]; in getCanonicalizationPatterns() 1999 rewriter.create<DimOp>(dimOp.getLoc(), expandShapeOp.getSrc(), srcDim); in fold() 1909 int64_t srcDim = expandShapeOp.getCorrespondingSourceDim(*dim); matchAndRewrite() local
|
/llvm-project/mlir/lib/Dialect/MemRef/IR/ |
H A D | MemRefOps.cpp | 2496 for (int64_t srcDim : group) in computeCollapsedType() 2498 groupSize * SaturatedInteger::wrap(srcType.getDimSize(srcDim)); in computeCollapsedType() 2475 for (int64_t srcDim : group) computeCollapsedType() local
|