diff --git a/pyproject.toml b/pyproject.toml index d2f1c650..72879b28 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -25,8 +25,8 @@ dependencies = [ "fastapi>=0.115.6", "uvicorn>=0.34.3", "kubernetes>=30.1.0", - "llama-stack==0.2.16", - "llama-stack-client==0.2.16", + "llama-stack==0.2.17", + "llama-stack-client==0.2.17", "rich>=14.0.0", "cachetools>=6.1.0", "prometheus-client>=0.22.1", diff --git a/src/client.py b/src/client.py index 93415008..04ddef41 100644 --- a/src/client.py +++ b/src/client.py @@ -4,7 +4,7 @@ from typing import Optional -from llama_stack.distribution.library_client import ( +from llama_stack import ( AsyncLlamaStackAsLibraryClient, # type: ignore LlamaStackAsLibraryClient, # type: ignore ) diff --git a/src/utils/common.py b/src/utils/common.py index 47ec67cb..e4a35a57 100644 --- a/src/utils/common.py +++ b/src/utils/common.py @@ -6,7 +6,7 @@ from logging import Logger from llama_stack_client import LlamaStackClient, AsyncLlamaStackClient -from llama_stack.distribution.library_client import ( +from llama_stack import ( AsyncLlamaStackAsLibraryClient, ) diff --git a/test.containerfile b/test.containerfile index b6b198a1..cd6f5785 100644 --- a/test.containerfile +++ b/test.containerfile @@ -8,7 +8,6 @@ ENV PATH="$PATH:/root/.local/bin" WORKDIR ${APP_ROOT} COPY run.yaml ./ - RUN microdnf install -y --nodocs --setopt=keepcache=0 --setopt=tsflags=nodocs \ python3.12 python3.12-devel python3.12-pip git tar @@ -17,7 +16,7 @@ RUN curl -LsSf https://astral.sh/uv/install.sh | sh RUN uv -h RUN uv venv && \ - uv pip install llama-stack==0.2.16 \ + uv pip install llama-stack==0.2.17 \ fastapi \ opentelemetry-sdk \ opentelemetry-exporter-otlp \ @@ -36,4 +35,4 @@ RUN uv venv && \ peft \ trl -CMD ["uv", "run", "llama", "stack", "run", "run.yaml"] \ No newline at end of file +CMD ["uv", "run", "llama", "stack", "run", "run.yaml"] diff --git a/uv.lock b/uv.lock index af7906ef..0843d848 100644 --- a/uv.lock +++ b/uv.lock @@ -902,8 +902,8 @@ requires-dist = [ { name = "cachetools", specifier = ">=6.1.0" }, { name = "fastapi", specifier = ">=0.115.6" }, { name = "kubernetes", specifier = ">=30.1.0" }, - { name = "llama-stack", specifier = "==0.2.16" }, - { name = "llama-stack-client", specifier = "==0.2.16" }, + { name = "llama-stack", specifier = "==0.2.17" }, + { name = "llama-stack-client", specifier = "==0.2.17" }, { name = "prometheus-client", specifier = ">=0.22.1" }, { name = "rich", specifier = ">=14.0.0" }, { name = "starlette", specifier = ">=0.47.1" }, @@ -954,7 +954,7 @@ wheels = [ [[package]] name = "llama-stack" -version = "0.2.16" +version = "0.2.17" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "aiohttp" }, @@ -984,14 +984,14 @@ dependencies = [ { name = "tiktoken" }, { name = "uvicorn" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/f9/76/3d73f4bfcf34ef7c703f3f3cfbe635f500b044c78fd33302199dce3ac698/llama_stack-0.2.16.tar.gz", hash = "sha256:e1ffb5400c85bf23b97f8f48028cb85061bb87a72b741faace1a174215f5de32", size = 3284603, upload-time = "2025-07-28T23:13:32.536Z" } +sdist = { url = "https://files.pythonhosted.org/packages/3b/43/ab4969bd47c9981efa713fa21c8c2ac534425831b265157248bfc710a508/llama_stack-0.2.17.tar.gz", hash = "sha256:f5e50f85a7483022bb86f9971bbe6f98274175e2e5584fb0a7df2d9ac43f5d57", size = 3301909, upload-time = "2025-08-05T01:43:04.059Z" } wheels = [ - { url = "https://files.pythonhosted.org/packages/24/36/35cc221603aa7c23efb41a88dd7c122af96740f50191d1da0ab9dc74056d/llama_stack-0.2.16-py3-none-any.whl", hash = "sha256:b9313acb150360467d7cccb54adb160b2fd585b7ab2505e88b5320c6bf766efe", size = 3603056, upload-time = "2025-07-28T23:13:30.731Z" }, + { url = "https://files.pythonhosted.org/packages/53/87/ebb127c7594cfecfa5b42db99a919a942e429a1571fd6c872890dfc9a6d9/llama_stack-0.2.17-py3-none-any.whl", hash = "sha256:b067575c26938932bf1a54246009b08b0082d5d5b511c5f14bd81a80167e046c", size = 3624746, upload-time = "2025-08-05T01:43:02.395Z" }, ] [[package]] name = "llama-stack-client" -version = "0.2.16" +version = "0.2.17" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "anyio" }, @@ -1010,9 +1010,9 @@ dependencies = [ { name = "tqdm" }, { name = "typing-extensions" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/db/28/74ae2faae9af51205587b33fcf2f99a8af090de7aa4122701f2f70f04233/llama_stack_client-0.2.16.tar.gz", hash = "sha256:24294acc6bf40e79900a62f4fa61009acb9af7028b198b12c0ba8adab25c2049", size = 257642, upload-time = "2025-07-28T23:13:22.793Z" } +sdist = { url = "https://files.pythonhosted.org/packages/c5/2a/bb2949d6a5c494d21da0c185d426e25eaa8016f8287b689249afc6c96fb5/llama_stack_client-0.2.17.tar.gz", hash = "sha256:1fe2070133c6356761e394fa346045e9b6b567d4c63157b9bc6be89b9a6e7a41", size = 257636, upload-time = "2025-08-05T01:42:55.911Z" } wheels = [ - { url = "https://files.pythonhosted.org/packages/30/ec/1874120a15b22f3a88d4e49700c870cc6540bc8c709a841db79a662d7949/llama_stack_client-0.2.16-py3-none-any.whl", hash = "sha256:5c0d13e6ac40143ce01cae4eec65fb39fe24e11f54b86afbd20f0033c38f83c0", size = 350329, upload-time = "2025-07-28T23:13:21.586Z" }, + { url = "https://files.pythonhosted.org/packages/81/fc/5eccc86b83c5ced3a3bca071d250a86ccafa4ff17546cf781deb7758ab74/llama_stack_client-0.2.17-py3-none-any.whl", hash = "sha256:336c32f8688700ff64717b8109f405dc87a990fbe310c2027ac9ed6d39d67d16", size = 350329, upload-time = "2025-08-05T01:42:54.381Z" }, ] [[package]]