diff --git a/engine/commands/run_cmd.cc b/engine/commands/run_cmd.cc index b4023cfd3..9ae71d85c 100644 --- a/engine/commands/run_cmd.cc +++ b/engine/commands/run_cmd.cc @@ -91,7 +91,8 @@ void RunCmd::Exec(bool chat_flag) { // Always start model if not llamacpp // If it is llamacpp, then check model status first { - if ((mc.engine.find("llamacpp") == std::string::npos) || + if ((mc.engine.find(kLlamaRepo) == std::string::npos && + mc.engine.find(kLlamaEngine) == std::string::npos) || !commands::ModelStatusCmd().IsLoaded(host_, port_, *model_id)) { if (!ModelStartCmd().Exec(host_, port_, *model_id)) { return; diff --git a/engine/utils/file_manager_utils.h b/engine/utils/file_manager_utils.h index c1e5e2d97..8d452e5f8 100644 --- a/engine/utils/file_manager_utils.h +++ b/engine/utils/file_manager_utils.h @@ -140,7 +140,7 @@ inline void CreateConfigFileIfNotExist() { CLI_LOG("Config file not found. Creating one at " + config_path.string()); auto defaultDataFolderPath = file_manager_utils::GetHomeDirectoryPath() / default_data_folder_name; - CTL_INF("Default data folder path: " + defaultDataFolderPath.string()); + CLI_LOG("Default data folder path: " + defaultDataFolderPath.string()); auto config = config_yaml_utils::CortexConfig{ .logFolderPath = defaultDataFolderPath.string(), @@ -188,7 +188,7 @@ inline std::filesystem::path GetCortexDataPath() { } if (!std::filesystem::exists(data_folder_path)) { - CTL_INF("Cortex home folder not found. Create one: " + + CLI_LOG("Cortex home folder not found. Create one: " + data_folder_path.string()); std::filesystem::create_directory(data_folder_path); }