diff --git a/src/app/endpoints/query.py b/src/app/endpoints/query.py index 758992ca..1b8819a1 100644 --- a/src/app/endpoints/query.py +++ b/src/app/endpoints/query.py @@ -85,7 +85,9 @@ def get_llama_stack_client( ) -> LlamaStackClient: if llama_stack_config.use_as_library_client is True: logger.info("Using Llama stack as library client") - client = LlamaStackAsLibraryClient("ollama") + client = LlamaStackAsLibraryClient( + llama_stack_config.library_client_config_path + ) client.initialize() return client else: diff --git a/src/models/config.py b/src/models/config.py index d3322ceb..5601465f 100644 --- a/src/models/config.py +++ b/src/models/config.py @@ -10,6 +10,7 @@ class LLamaStackConfiguration(BaseModel): url: Optional[str] = None api_key: Optional[str] = None use_as_library_client: Optional[bool] = None + library_client_config_path: Optional[str] = None chat_completion_mode: bool = False @model_validator(mode="after") @@ -25,6 +26,11 @@ def check_llama_stack_model(self) -> Self: ) if self.use_as_library_client is None: self.use_as_library_client = False + if self.use_as_library_client: + if self.library_client_config_path is None: + raise ValueError( + "LLama stack library client mode is enabled but a configuration file path is not specified" + ) return self