Home
last modified time | relevance | path

Searched refs:TensorSpec (Results 1 – 25 of 33) sorted by relevance

12

/openbsd-src/gnu/llvm/llvm/lib/Analysis/
H A DModelUnderTrainingRunner.cpp25 TensorSpec Spec;
60 if (auto TensorSpec = getTensorSpecFromJSON(Ctx, *SpecPart)) in loadOutputSpecs() local
62 if (!TensorSpec->isElementType<int64_t>() && in loadOutputSpecs()
63 !TensorSpec->isElementType<int32_t>() && in loadOutputSpecs()
64 !TensorSpec->isElementType<float>()) { in loadOutputSpecs()
68 TensorSpec->name()); in loadOutputSpecs()
71 Ret.push_back({*TensorSpec, LoggingName->str()}); in loadOutputSpecs()
95 const std::vector<TensorSpec> &InputSpecs, in ModelUnderTrainingRunner()
96 const std::vector<TensorSpec> &OutputSpecs, in ModelUnderTrainingRunner()
97 const std::vector<TensorSpec> &ExtraOutputsForLogging) in ModelUnderTrainingRunner()
[all …]
H A DTensorSpec.cpp31 template <> TensorType TensorSpec::getDataType<T>() { return TensorType::E; }
48 void TensorSpec::toJSON(json::OStream &OS) const { in toJSON()
60 TensorSpec::TensorSpec(const std::string &Name, int Port, TensorType Type, in TensorSpec() function in llvm::TensorSpec
67 std::optional<TensorSpec> getTensorSpecFromJSON(LLVMContext &Ctx, in getTensorSpecFromJSON()
70 [&](const llvm::Twine &Message) -> std::optional<TensorSpec> { in getTensorSpecFromJSON()
99 return TensorSpec::createSpec<T>(TensorName, TensorShape, TensorPort); in getTensorSpecFromJSON()
H A DTFLiteUtils.cpp56 const std::vector<TensorSpec> &InputSpecs,
57 const std::vector<TensorSpec> &OutputSpecs,
95 const TensorSpec &Spec);
101 StringRef SavedModelPath, const std::vector<TensorSpec> &InputSpecs, in TFModelEvaluatorImpl()
102 const std::vector<TensorSpec> &OutputSpecs, const char *Tags = "serve") in TFModelEvaluatorImpl()
185 const std::vector<TensorSpec> &InputSpecs, in TFModelEvaluator()
186 const std::vector<TensorSpec> &OutputSpecs, in TFModelEvaluator()
197 const TensorSpec &Spec) { in checkReportAndInvalidate()
H A DInlineSizeEstimatorAnalysis.cpp222 std::vector<TensorSpec> InputSpecs{TensorSpec::createSpec<int32_t>( in InlineSizeEstimatorAnalysis()
226 std::vector<TensorSpec> OutputSpecs{ in InlineSizeEstimatorAnalysis()
227 TensorSpec::createSpec<float>("StatefulPartitionedCall", {1})}; in InlineSizeEstimatorAnalysis()
H A DTrainingLogger.cpp83 const std::vector<TensorSpec> &FeatureSpecs, in Logger()
84 const TensorSpec &RewardSpec, bool IncludeReward) in Logger()
H A DNoInferenceModelRunner.cpp18 LLVMContext &Ctx, const std::vector<TensorSpec> &Inputs) in NoInferenceModelRunner()
H A DDevelopmentModeInlineAdvisor.cpp47 - "tensor_spec, followed by the TensorSpec description of the
H A DMLInlineAdvisor.cpp62 const std::array<TensorSpec, NumberOfFeatures> llvm::FeatureMap{
63 #define POPULATE_NAMES(_, NAME) TensorSpec::createSpec<int64_t>(NAME, {1} ),
69 #define POPULATE_NAMES(_, NAME, __) TensorSpec::createSpec<int64_t>(NAME, {1} ),
H A DCMakeLists.txt136 TensorSpec.cpp
/openbsd-src/gnu/llvm/llvm/include/llvm/Analysis/
H A DTensorSpec.h54 class TensorSpec final {
57 static TensorSpec createSpec(const std::string &Name,
60 return TensorSpec(Name, Port, getDataType<T>(), sizeof(T), Shape);
68 bool operator==(const TensorSpec &Other) const {
73 bool operator!=(const TensorSpec &Other) const { return !(*this == Other); }
86 TensorSpec(const std::string &NewName, const TensorSpec &Other) in TensorSpec() function
87 : TensorSpec(NewName, Other.Port, Other.Type, Other.ElementSize, in TensorSpec()
93 TensorSpec(const std::string &Name, int Port, TensorType Type,
113 std::optional<TensorSpec> getTensorSpecFromJSON(LLVMContext &Ctx,
117 template <> TensorType TensorSpec::getDataType<T>();
H A DModelUnderTrainingRunner.h37 const std::vector<TensorSpec> &extraOutputsForLoggingSpecs() const { in extraOutputsForLoggingSpecs()
56 const std::vector<TensorSpec> &InputSpecs,
61 const std::vector<TensorSpec> &InputSpecs,
62 const std::vector<TensorSpec> &OutputSpecs,
63 const std::vector<TensorSpec> &ExtraOutputsForLogging = {});
69 const std::vector<TensorSpec> OutputSpecs;
70 const std::vector<TensorSpec> ExtraOutputsForLogging;
H A DNoInferenceModelRunner.h23 const std::vector<TensorSpec> &Inputs);
H A DMLModelRunner.h60 void setUpBufferForTensor(size_t Index, const TensorSpec &Spec, in setUpBufferForTensor()
H A DInlineModelFeatureMaps.h132 extern const std::array<TensorSpec, NumberOfFeatures> FeatureMap;
/openbsd-src/gnu/llvm/llvm/lib/CodeGen/
H A DMLRegallocPriorityAdvisor.cpp99 TensorSpec::createSpec<type>(#name, shape),
101 static const std::vector<TensorSpec> InputFeatures{
144 static const TensorSpec Output =
145 TensorSpec::createSpec<float>(DecisionName, {1});
146 static const TensorSpec Reward = TensorSpec::createSpec<float>("reward", {1});
149 TensorSpec::createSpec<type>(std::string("action_") + #name, shape),
151 static const std::vector<TensorSpec> TrainingInputFeatures{
153 TensorSpec::createSpec<float>("action_discount", {1}),
154 TensorSpec::createSpec<int32_t>("action_step_type", {1}),
155 TensorSpec::createSpec<float>("action_reward", {1})}};
[all …]
H A DMLRegallocEvictAdvisor.cpp351 TensorSpec::createSpec<type>(#name, shape),
375 std::vector<TensorSpec> InputFeatures;
401 static const TensorSpec Output =
402 TensorSpec::createSpec<int64_t>(DecisionName, {1});
403 static const TensorSpec Reward = TensorSpec::createSpec<float>("reward", {1});
410 TensorSpec::createSpec<type>(std::string("action_") + #name, shape),
442 TensorSpec::createSpec<float>("action_discount", {1}), in DevelopmentModeEvictionAdvisorAnalysis()
443 TensorSpec::createSpec<int32_t>("action_step_type", {1}), in DevelopmentModeEvictionAdvisorAnalysis()
444 TensorSpec::createSpec<float>("action_reward", {1})}; in DevelopmentModeEvictionAdvisorAnalysis()
449 TensorSpec::createSpec<float>("action_discount", {1}), in DevelopmentModeEvictionAdvisorAnalysis()
[all …]
/openbsd-src/gnu/llvm/llvm/include/llvm/Analysis/Utils/
H A DTrainingLogger.h93 const std::vector<TensorSpec> FeatureSpecs;
94 const TensorSpec RewardSpec;
100 void writeTensor(const TensorSpec &Spec, const char *RawData) { in writeTensor()
113 const std::vector<TensorSpec> &FeatureSpecs,
114 const TensorSpec &RewardSpec, bool IncludeReward);
H A DTFUtils.h77 const std::vector<TensorSpec> &InputSpecs,
78 const std::vector<TensorSpec> &OutputSpecs,
/openbsd-src/gnu/llvm/llvm/lib/Analysis/models/
H A Dlog_reader.py27 class TensorSpec: class
41 return TensorSpec(
50 def __init__(self, spec: TensorSpec, buffer: bytes): argument
57 def spec(self) -> TensorSpec:
69 def read_tensor(fs: typing.BinaryIO, ts: TensorSpec) -> TensorValue: argument
82 tensor_specs = [TensorSpec.from_dict(ts) for ts in header['features']]
83 score_spec = TensorSpec.from_dict(
H A Dgen-regalloc-priority-test-model.py39 (key, tf.TensorSpec(dtype=tf.int64, shape=(), name=key))
43 tf.TensorSpec(dtype=tf.float32, shape=(), name=key))
46 dict((key, tf.TensorSpec(dtype=tf.float32, shape=(), name=key))
49 dict((key, tf.TensorSpec(dtype=tf.int32, shape=(), name=key))
H A Dgen-inline-oz-test-model.py37 tf.TensorSpec(dtype=tf.int64, shape=(), name=key) for key in [
79 tf.TensorSpec(dtype=tf.float32, shape=(), name=key)
85 tf.TensorSpec(dtype=tf.int32, shape=(), name=key)
H A Dgen-regalloc-eviction-test-model.py32 (key, tf.TensorSpec(dtype=tf.int64, shape=(NUM_REGISTERS), name=key))
/openbsd-src/gnu/usr.bin/clang/libLLVMAnalysis/
H A DMakefile117 TensorSpec.cpp \
/openbsd-src/gnu/llvm/llvm/utils/gn/secondary/llvm/lib/Analysis/
H A DBUILD.gn126 "TensorSpec.cpp",
/openbsd-src/distrib/sets/lists/comp/
H A Dclang.amd641191 ./usr/include/llvm/Analysis/TensorSpec.h

12