diff --git a/llvm/lib/Analysis/InteractiveModelRunner.cpp b/llvm/lib/Analysis/InteractiveModelRunner.cpp index 2e9a906964b14..a0f42fb59190c 100644 --- a/llvm/lib/Analysis/InteractiveModelRunner.cpp +++ b/llvm/lib/Analysis/InteractiveModelRunner.cpp @@ -18,18 +18,10 @@ using namespace llvm; -#define _IMR_CL_VALS(T, N) clEnumValN(TensorType::N, #T, #T), - -static cl::opt DebugReply( - "interactive-model-runner-echo-type", cl::init(TensorType::Invalid), - cl::Hidden, +static cl::opt DebugReply( + "interactive-model-runner-echo-reply", cl::init(false), cl::Hidden, cl::desc("The InteractiveModelRunner will echo back to stderr " - "the data received " - "from the host as the specified type (for debugging purposes)."), - cl::values(SUPPORTED_TENSOR_TYPES(_IMR_CL_VALS) - clEnumValN(TensorType::Invalid, "disable", "Don't echo"))); - -#undef _IMR_CL_VALS + "the data received from the host (for debugging purposes).")); InteractiveModelRunner::InteractiveModelRunner( LLVMContext &Ctx, const std::vector &Inputs, @@ -75,15 +67,15 @@ void *InteractiveModelRunner::evaluateUntyped() { const size_t Limit = OutputBuffer.size(); while (InsPoint < Limit) { auto ReadOrErr = ::sys::fs::readNativeFile( - sys::fs::convertFDToNativeFile(Inbound), - {Buff + InsPoint, OutputBuffer.size() - InsPoint}); + Inbound, {Buff + InsPoint, OutputBuffer.size() - InsPoint}); if (ReadOrErr.takeError()) { Ctx.emitError("Failed reading from inbound file"); break; } InsPoint += *ReadOrErr; } - if (DebugReply != TensorType::Invalid) - dbgs() << tensorValueToString(OutputBuffer.data(), OutputSpec); + if (DebugReply) + dbgs() << OutputSpec.name() << ": " + << tensorValueToString(OutputBuffer.data(), OutputSpec) << "\n"; return OutputBuffer.data(); }