| /openbsd-src/gnu/llvm/llvm/lib/Analysis/ |
| H A D | ModelUnderTrainingRunner.cpp | 25 TensorSpec Spec; 60 if (auto TensorSpec = getTensorSpecFromJSON(Ctx, *SpecPart)) in loadOutputSpecs() local 62 if (!TensorSpec->isElementType<int64_t>() && in loadOutputSpecs() 63 !TensorSpec->isElementType<int32_t>() && in loadOutputSpecs() 64 !TensorSpec->isElementType<float>()) { in loadOutputSpecs() 68 TensorSpec->name()); in loadOutputSpecs() 71 Ret.push_back({*TensorSpec, LoggingName->str()}); in loadOutputSpecs() 95 const std::vector<TensorSpec> &InputSpecs, in ModelUnderTrainingRunner() 96 const std::vector<TensorSpec> &OutputSpecs, in ModelUnderTrainingRunner() 97 const std::vector<TensorSpec> &ExtraOutputsForLogging) in ModelUnderTrainingRunner() [all …]
|
| H A D | TensorSpec.cpp | 31 template <> TensorType TensorSpec::getDataType<T>() { return TensorType::E; } 48 void TensorSpec::toJSON(json::OStream &OS) const { in toJSON() 60 TensorSpec::TensorSpec(const std::string &Name, int Port, TensorType Type, in TensorSpec() function in llvm::TensorSpec 67 std::optional<TensorSpec> getTensorSpecFromJSON(LLVMContext &Ctx, in getTensorSpecFromJSON() 70 [&](const llvm::Twine &Message) -> std::optional<TensorSpec> { in getTensorSpecFromJSON() 99 return TensorSpec::createSpec<T>(TensorName, TensorShape, TensorPort); in getTensorSpecFromJSON()
|
| H A D | TFLiteUtils.cpp | 56 const std::vector<TensorSpec> &InputSpecs, 57 const std::vector<TensorSpec> &OutputSpecs, 95 const TensorSpec &Spec); 101 StringRef SavedModelPath, const std::vector<TensorSpec> &InputSpecs, in TFModelEvaluatorImpl() 102 const std::vector<TensorSpec> &OutputSpecs, const char *Tags = "serve") in TFModelEvaluatorImpl() 185 const std::vector<TensorSpec> &InputSpecs, in TFModelEvaluator() 186 const std::vector<TensorSpec> &OutputSpecs, in TFModelEvaluator() 197 const TensorSpec &Spec) { in checkReportAndInvalidate()
|
| H A D | InlineSizeEstimatorAnalysis.cpp | 222 std::vector<TensorSpec> InputSpecs{TensorSpec::createSpec<int32_t>( in InlineSizeEstimatorAnalysis() 226 std::vector<TensorSpec> OutputSpecs{ in InlineSizeEstimatorAnalysis() 227 TensorSpec::createSpec<float>("StatefulPartitionedCall", {1})}; in InlineSizeEstimatorAnalysis()
|
| H A D | TrainingLogger.cpp | 83 const std::vector<TensorSpec> &FeatureSpecs, in Logger() 84 const TensorSpec &RewardSpec, bool IncludeReward) in Logger()
|
| H A D | NoInferenceModelRunner.cpp | 18 LLVMContext &Ctx, const std::vector<TensorSpec> &Inputs) in NoInferenceModelRunner()
|
| H A D | DevelopmentModeInlineAdvisor.cpp | 47 - "tensor_spec, followed by the TensorSpec description of the
|
| H A D | MLInlineAdvisor.cpp | 62 const std::array<TensorSpec, NumberOfFeatures> llvm::FeatureMap{ 63 #define POPULATE_NAMES(_, NAME) TensorSpec::createSpec<int64_t>(NAME, {1} ), 69 #define POPULATE_NAMES(_, NAME, __) TensorSpec::createSpec<int64_t>(NAME, {1} ),
|
| H A D | CMakeLists.txt | 136 TensorSpec.cpp
|
| /openbsd-src/gnu/llvm/llvm/include/llvm/Analysis/ |
| H A D | TensorSpec.h | 54 class TensorSpec final { 57 static TensorSpec createSpec(const std::string &Name, 60 return TensorSpec(Name, Port, getDataType<T>(), sizeof(T), Shape); 68 bool operator==(const TensorSpec &Other) const { 73 bool operator!=(const TensorSpec &Other) const { return !(*this == Other); } 86 TensorSpec(const std::string &NewName, const TensorSpec &Other) in TensorSpec() function 87 : TensorSpec(NewName, Other.Port, Other.Type, Other.ElementSize, in TensorSpec() 93 TensorSpec(const std::string &Name, int Port, TensorType Type, 113 std::optional<TensorSpec> getTensorSpecFromJSON(LLVMContext &Ctx, 117 template <> TensorType TensorSpec::getDataType<T>();
|
| H A D | ModelUnderTrainingRunner.h | 37 const std::vector<TensorSpec> &extraOutputsForLoggingSpecs() const { in extraOutputsForLoggingSpecs() 56 const std::vector<TensorSpec> &InputSpecs, 61 const std::vector<TensorSpec> &InputSpecs, 62 const std::vector<TensorSpec> &OutputSpecs, 63 const std::vector<TensorSpec> &ExtraOutputsForLogging = {}); 69 const std::vector<TensorSpec> OutputSpecs; 70 const std::vector<TensorSpec> ExtraOutputsForLogging;
|
| H A D | NoInferenceModelRunner.h | 23 const std::vector<TensorSpec> &Inputs);
|
| H A D | MLModelRunner.h | 60 void setUpBufferForTensor(size_t Index, const TensorSpec &Spec, in setUpBufferForTensor()
|
| H A D | InlineModelFeatureMaps.h | 132 extern const std::array<TensorSpec, NumberOfFeatures> FeatureMap;
|
| /openbsd-src/gnu/llvm/llvm/lib/CodeGen/ |
| H A D | MLRegallocPriorityAdvisor.cpp | 99 TensorSpec::createSpec<type>(#name, shape), 101 static const std::vector<TensorSpec> InputFeatures{ 144 static const TensorSpec Output = 145 TensorSpec::createSpec<float>(DecisionName, {1}); 146 static const TensorSpec Reward = TensorSpec::createSpec<float>("reward", {1}); 149 TensorSpec::createSpec<type>(std::string("action_") + #name, shape), 151 static const std::vector<TensorSpec> TrainingInputFeatures{ 153 TensorSpec::createSpec<float>("action_discount", {1}), 154 TensorSpec::createSpec<int32_t>("action_step_type", {1}), 155 TensorSpec::createSpec<float>("action_reward", {1})}}; [all …]
|
| H A D | MLRegallocEvictAdvisor.cpp | 351 TensorSpec::createSpec<type>(#name, shape), 375 std::vector<TensorSpec> InputFeatures; 401 static const TensorSpec Output = 402 TensorSpec::createSpec<int64_t>(DecisionName, {1}); 403 static const TensorSpec Reward = TensorSpec::createSpec<float>("reward", {1}); 410 TensorSpec::createSpec<type>(std::string("action_") + #name, shape), 442 TensorSpec::createSpec<float>("action_discount", {1}), in DevelopmentModeEvictionAdvisorAnalysis() 443 TensorSpec::createSpec<int32_t>("action_step_type", {1}), in DevelopmentModeEvictionAdvisorAnalysis() 444 TensorSpec::createSpec<float>("action_reward", {1})}; in DevelopmentModeEvictionAdvisorAnalysis() 449 TensorSpec::createSpec<float>("action_discount", {1}), in DevelopmentModeEvictionAdvisorAnalysis() [all …]
|
| /openbsd-src/gnu/llvm/llvm/include/llvm/Analysis/Utils/ |
| H A D | TrainingLogger.h | 93 const std::vector<TensorSpec> FeatureSpecs; 94 const TensorSpec RewardSpec; 100 void writeTensor(const TensorSpec &Spec, const char *RawData) { in writeTensor() 113 const std::vector<TensorSpec> &FeatureSpecs, 114 const TensorSpec &RewardSpec, bool IncludeReward);
|
| H A D | TFUtils.h | 77 const std::vector<TensorSpec> &InputSpecs, 78 const std::vector<TensorSpec> &OutputSpecs,
|
| /openbsd-src/gnu/llvm/llvm/lib/Analysis/models/ |
| H A D | log_reader.py | 27 class TensorSpec: class 41 return TensorSpec( 50 def __init__(self, spec: TensorSpec, buffer: bytes): argument 57 def spec(self) -> TensorSpec: 69 def read_tensor(fs: typing.BinaryIO, ts: TensorSpec) -> TensorValue: argument 82 tensor_specs = [TensorSpec.from_dict(ts) for ts in header['features']] 83 score_spec = TensorSpec.from_dict(
|
| H A D | gen-regalloc-priority-test-model.py | 39 (key, tf.TensorSpec(dtype=tf.int64, shape=(), name=key)) 43 tf.TensorSpec(dtype=tf.float32, shape=(), name=key)) 46 dict((key, tf.TensorSpec(dtype=tf.float32, shape=(), name=key)) 49 dict((key, tf.TensorSpec(dtype=tf.int32, shape=(), name=key))
|
| H A D | gen-inline-oz-test-model.py | 37 tf.TensorSpec(dtype=tf.int64, shape=(), name=key) for key in [ 79 tf.TensorSpec(dtype=tf.float32, shape=(), name=key) 85 tf.TensorSpec(dtype=tf.int32, shape=(), name=key)
|
| H A D | gen-regalloc-eviction-test-model.py | 32 (key, tf.TensorSpec(dtype=tf.int64, shape=(NUM_REGISTERS), name=key))
|
| /openbsd-src/gnu/usr.bin/clang/libLLVMAnalysis/ |
| H A D | Makefile | 117 TensorSpec.cpp \
|
| /openbsd-src/gnu/llvm/llvm/utils/gn/secondary/llvm/lib/Analysis/ |
| H A D | BUILD.gn | 126 "TensorSpec.cpp",
|
| /openbsd-src/distrib/sets/lists/comp/ |
| H A D | clang.amd64 | 1191 ./usr/include/llvm/Analysis/TensorSpec.h
|