From 03a2e2b52d946a01b76b59382bf1e1f544c646b7 Mon Sep 17 00:00:00 2001 From: Wu Jian Ping Date: Fri, 26 Jan 2024 16:04:51 +0800 Subject: [PATCH] add checking model loaded first in embedding call --- controllers/llamaCPP.cc | 11 +++++++++++ 1 file changed, 11 insertions(+) diff --git a/controllers/llamaCPP.cc b/controllers/llamaCPP.cc index faf56b2bd..19be073ae 100644 --- a/controllers/llamaCPP.cc +++ b/controllers/llamaCPP.cc @@ -143,6 +143,7 @@ void llamaCPP::chatCompletion( auto resp = nitro_utils::nitroHttpJsonResponse(jsonResp); resp->setStatusCode(drogon::k409Conflict); callback(resp); + return; } const auto &jsonBody = req->getJsonObject(); @@ -401,6 +402,16 @@ void llamaCPP::chatCompletion( void llamaCPP::embedding( const HttpRequestPtr &req, std::function &&callback) { + if (!llama.model_loaded_external) { + Json::Value jsonResp; + jsonResp["message"] = + "Model has not been loaded, please load model into nitro"; + auto resp = nitro_utils::nitroHttpJsonResponse(jsonResp); + resp->setStatusCode(drogon::k409Conflict); + callback(resp); + return; + } + const auto &jsonBody = req->getJsonObject(); Json::Value responseData(Json::arrayValue);