xref: /llvm-project/llvm/lib/Analysis/InteractiveModelRunner.cpp (revision 9b06f756b2438684e0a3357d3f6fd12b50a2092b)
15b8dc7c8SMircea Trofin //===- InteractiveModelRunner.cpp - noop ML model runner   ----------------===//
25b8dc7c8SMircea Trofin //
35b8dc7c8SMircea Trofin // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
45b8dc7c8SMircea Trofin // See https://llvm.org/LICENSE.txt for license information.
55b8dc7c8SMircea Trofin // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
65b8dc7c8SMircea Trofin //
75b8dc7c8SMircea Trofin //===----------------------------------------------------------------------===//
85b8dc7c8SMircea Trofin //
95b8dc7c8SMircea Trofin // A runner that communicates with an external agent via 2 file descriptors.
105b8dc7c8SMircea Trofin //===----------------------------------------------------------------------===//
115b8dc7c8SMircea Trofin #include "llvm/Analysis/InteractiveModelRunner.h"
125b8dc7c8SMircea Trofin #include "llvm/Analysis/MLModelRunner.h"
135b8dc7c8SMircea Trofin #include "llvm/Analysis/TensorSpec.h"
145b8dc7c8SMircea Trofin #include "llvm/Support/CommandLine.h"
155b8dc7c8SMircea Trofin #include "llvm/Support/ErrorHandling.h"
1683051c5aSMircea Trofin #include "llvm/Support/FileSystem.h"
175b8dc7c8SMircea Trofin #include "llvm/Support/raw_ostream.h"
185b8dc7c8SMircea Trofin 
195b8dc7c8SMircea Trofin using namespace llvm;
205b8dc7c8SMircea Trofin 
21*9b06f756SMircea Trofin static cl::opt<bool> DebugReply(
22*9b06f756SMircea Trofin     "interactive-model-runner-echo-reply", cl::init(false), cl::Hidden,
235b8dc7c8SMircea Trofin     cl::desc("The InteractiveModelRunner will echo back to stderr "
24*9b06f756SMircea Trofin              "the data received from the host (for debugging purposes)."));
255b8dc7c8SMircea Trofin 
InteractiveModelRunner(LLVMContext & Ctx,const std::vector<TensorSpec> & Inputs,const TensorSpec & Advice,StringRef OutboundName,StringRef InboundName)265b8dc7c8SMircea Trofin InteractiveModelRunner::InteractiveModelRunner(
275b8dc7c8SMircea Trofin     LLVMContext &Ctx, const std::vector<TensorSpec> &Inputs,
285b8dc7c8SMircea Trofin     const TensorSpec &Advice, StringRef OutboundName, StringRef InboundName)
295b8dc7c8SMircea Trofin     : MLModelRunner(Ctx, MLModelRunner::Kind::Interactive, Inputs.size()),
3083051c5aSMircea Trofin       InputSpecs(Inputs), OutputSpec(Advice),
3183051c5aSMircea Trofin       InEC(sys::fs::openFileForRead(InboundName, Inbound)),
3283051c5aSMircea Trofin       OutputBuffer(OutputSpec.getTotalTensorBufferSize()) {
335b8dc7c8SMircea Trofin   if (InEC) {
345b8dc7c8SMircea Trofin     Ctx.emitError("Cannot open inbound file: " + InEC.message());
355b8dc7c8SMircea Trofin     return;
365b8dc7c8SMircea Trofin   }
3783051c5aSMircea Trofin   {
3883051c5aSMircea Trofin     auto OutStream = std::make_unique<raw_fd_ostream>(OutboundName, OutEC);
395b8dc7c8SMircea Trofin     if (OutEC) {
405b8dc7c8SMircea Trofin       Ctx.emitError("Cannot open outbound file: " + OutEC.message());
415b8dc7c8SMircea Trofin       return;
425b8dc7c8SMircea Trofin     }
4383051c5aSMircea Trofin     Log = std::make_unique<Logger>(std::move(OutStream), InputSpecs, Advice,
4483051c5aSMircea Trofin                                    /*IncludeReward=*/false, Advice);
4583051c5aSMircea Trofin   }
465b8dc7c8SMircea Trofin   // Just like in the no inference case, this will allocate an appropriately
475b8dc7c8SMircea Trofin   // sized buffer.
485b8dc7c8SMircea Trofin   for (size_t I = 0; I < InputSpecs.size(); ++I)
495b8dc7c8SMircea Trofin     setUpBufferForTensor(I, InputSpecs[I], nullptr);
5083051c5aSMircea Trofin   Log->flush();
5183051c5aSMircea Trofin }
5283051c5aSMircea Trofin 
~InteractiveModelRunner()5383051c5aSMircea Trofin InteractiveModelRunner::~InteractiveModelRunner() {
54795910c2SMircea Trofin   sys::fs::file_t FDAsOSHandle = sys::fs::convertFDToNativeFile(Inbound);
55795910c2SMircea Trofin   sys::fs::closeFile(FDAsOSHandle);
565b8dc7c8SMircea Trofin }
575b8dc7c8SMircea Trofin 
evaluateUntyped()585b8dc7c8SMircea Trofin void *InteractiveModelRunner::evaluateUntyped() {
5983051c5aSMircea Trofin   Log->startObservation();
605b8dc7c8SMircea Trofin   for (size_t I = 0; I < InputSpecs.size(); ++I)
6183051c5aSMircea Trofin     Log->logTensorValue(I, reinterpret_cast<const char *>(getTensorUntyped(I)));
6283051c5aSMircea Trofin   Log->endObservation();
6383051c5aSMircea Trofin   Log->flush();
645b8dc7c8SMircea Trofin 
655b8dc7c8SMircea Trofin   size_t InsPoint = 0;
665b8dc7c8SMircea Trofin   char *Buff = OutputBuffer.data();
675b8dc7c8SMircea Trofin   const size_t Limit = OutputBuffer.size();
685b8dc7c8SMircea Trofin   while (InsPoint < Limit) {
6983051c5aSMircea Trofin     auto ReadOrErr = ::sys::fs::readNativeFile(
70735f117fSMircea Trofin         sys::fs::convertFDToNativeFile(Inbound),
71735f117fSMircea Trofin         {Buff + InsPoint, OutputBuffer.size() - InsPoint});
7283051c5aSMircea Trofin     if (ReadOrErr.takeError()) {
735b8dc7c8SMircea Trofin       Ctx.emitError("Failed reading from inbound file");
745b8dc7c8SMircea Trofin       break;
755b8dc7c8SMircea Trofin     }
7683051c5aSMircea Trofin     InsPoint += *ReadOrErr;
775b8dc7c8SMircea Trofin   }
78*9b06f756SMircea Trofin   if (DebugReply)
79*9b06f756SMircea Trofin     dbgs() << OutputSpec.name() << ": "
80*9b06f756SMircea Trofin            << tensorValueToString(OutputBuffer.data(), OutputSpec) << "\n";
815b8dc7c8SMircea Trofin   return OutputBuffer.data();
825b8dc7c8SMircea Trofin }
83