diff --git a/controllers/llamaCPP.cc b/controllers/llamaCPP.cc index 881b9632c..f264a25e2 100644 --- a/controllers/llamaCPP.cc +++ b/controllers/llamaCPP.cc @@ -149,6 +149,15 @@ void llamaCPP::chatCompletion( const HttpRequestPtr &req, std::function &&callback) { + if (!model_loaded) { + Json::Value jsonResp; + jsonResp["message"] = + "Model has not been loaded, please load model into nitro"; + auto resp = nitro_utils::nitroHttpJsonResponse(jsonResp); + resp->setStatusCode(drogon::k409Conflict); + callback(resp); + } + const auto &jsonBody = req->getJsonObject(); std::string formatted_output = pre_prompt; @@ -338,6 +347,16 @@ void llamaCPP::loadModel( const HttpRequestPtr &req, std::function &&callback) { + if (model_loaded) { + LOG_INFO << "model loaded"; + Json::Value jsonResp; + jsonResp["message"] = "Model already loaded"; + auto resp = nitro_utils::nitroHttpJsonResponse(jsonResp); + resp->setStatusCode(drogon::k409Conflict); + callback(resp); + return; + } + const auto &jsonBody = req->getJsonObject(); gpt_params params;