From 78b81237164034c48ea3932373d0b9c8e5072d61 Mon Sep 17 00:00:00 2001 From: Pavel Tisnovsky Date: Wed, 6 Aug 2025 08:48:31 +0200 Subject: [PATCH 1/5] LCORE-336: bump-up Llama Stack to 0.2.17 --- pyproject.toml | 4 ++-- uv.lock | 16 ++++++++-------- 2 files changed, 10 insertions(+), 10 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index d2f1c650..72879b28 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -25,8 +25,8 @@ dependencies = [ "fastapi>=0.115.6", "uvicorn>=0.34.3", "kubernetes>=30.1.0", - "llama-stack==0.2.16", - "llama-stack-client==0.2.16", + "llama-stack==0.2.17", + "llama-stack-client==0.2.17", "rich>=14.0.0", "cachetools>=6.1.0", "prometheus-client>=0.22.1", diff --git a/uv.lock b/uv.lock index af7906ef..0843d848 100644 --- a/uv.lock +++ b/uv.lock @@ -902,8 +902,8 @@ requires-dist = [ { name = "cachetools", specifier = ">=6.1.0" }, { name = "fastapi", specifier = ">=0.115.6" }, { name = "kubernetes", specifier = ">=30.1.0" }, - { name = "llama-stack", specifier = "==0.2.16" }, - { name = "llama-stack-client", specifier = "==0.2.16" }, + { name = "llama-stack", specifier = "==0.2.17" }, + { name = "llama-stack-client", specifier = "==0.2.17" }, { name = "prometheus-client", specifier = ">=0.22.1" }, { name = "rich", specifier = ">=14.0.0" }, { name = "starlette", specifier = ">=0.47.1" }, @@ -954,7 +954,7 @@ wheels = [ [[package]] name = "llama-stack" -version = "0.2.16" +version = "0.2.17" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "aiohttp" }, @@ -984,14 +984,14 @@ dependencies = [ { name = "tiktoken" }, { name = "uvicorn" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/f9/76/3d73f4bfcf34ef7c703f3f3cfbe635f500b044c78fd33302199dce3ac698/llama_stack-0.2.16.tar.gz", hash = "sha256:e1ffb5400c85bf23b97f8f48028cb85061bb87a72b741faace1a174215f5de32", size = 3284603, upload-time = "2025-07-28T23:13:32.536Z" } +sdist = { url = "https://files.pythonhosted.org/packages/3b/43/ab4969bd47c9981efa713fa21c8c2ac534425831b265157248bfc710a508/llama_stack-0.2.17.tar.gz", hash = "sha256:f5e50f85a7483022bb86f9971bbe6f98274175e2e5584fb0a7df2d9ac43f5d57", size = 3301909, upload-time = "2025-08-05T01:43:04.059Z" } wheels = [ - { url = "https://files.pythonhosted.org/packages/24/36/35cc221603aa7c23efb41a88dd7c122af96740f50191d1da0ab9dc74056d/llama_stack-0.2.16-py3-none-any.whl", hash = "sha256:b9313acb150360467d7cccb54adb160b2fd585b7ab2505e88b5320c6bf766efe", size = 3603056, upload-time = "2025-07-28T23:13:30.731Z" }, + { url = "https://files.pythonhosted.org/packages/53/87/ebb127c7594cfecfa5b42db99a919a942e429a1571fd6c872890dfc9a6d9/llama_stack-0.2.17-py3-none-any.whl", hash = "sha256:b067575c26938932bf1a54246009b08b0082d5d5b511c5f14bd81a80167e046c", size = 3624746, upload-time = "2025-08-05T01:43:02.395Z" }, ] [[package]] name = "llama-stack-client" -version = "0.2.16" +version = "0.2.17" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "anyio" }, @@ -1010,9 +1010,9 @@ dependencies = [ { name = "tqdm" }, { name = "typing-extensions" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/db/28/74ae2faae9af51205587b33fcf2f99a8af090de7aa4122701f2f70f04233/llama_stack_client-0.2.16.tar.gz", hash = "sha256:24294acc6bf40e79900a62f4fa61009acb9af7028b198b12c0ba8adab25c2049", size = 257642, upload-time = "2025-07-28T23:13:22.793Z" } +sdist = { url = "https://files.pythonhosted.org/packages/c5/2a/bb2949d6a5c494d21da0c185d426e25eaa8016f8287b689249afc6c96fb5/llama_stack_client-0.2.17.tar.gz", hash = "sha256:1fe2070133c6356761e394fa346045e9b6b567d4c63157b9bc6be89b9a6e7a41", size = 257636, upload-time = "2025-08-05T01:42:55.911Z" } wheels = [ - { url = "https://files.pythonhosted.org/packages/30/ec/1874120a15b22f3a88d4e49700c870cc6540bc8c709a841db79a662d7949/llama_stack_client-0.2.16-py3-none-any.whl", hash = "sha256:5c0d13e6ac40143ce01cae4eec65fb39fe24e11f54b86afbd20f0033c38f83c0", size = 350329, upload-time = "2025-07-28T23:13:21.586Z" }, + { url = "https://files.pythonhosted.org/packages/81/fc/5eccc86b83c5ced3a3bca071d250a86ccafa4ff17546cf781deb7758ab74/llama_stack_client-0.2.17-py3-none-any.whl", hash = "sha256:336c32f8688700ff64717b8109f405dc87a990fbe310c2027ac9ed6d39d67d16", size = 350329, upload-time = "2025-08-05T01:42:54.381Z" }, ] [[package]] From c616e3617d806c05dae4c2820ff352bb78d4148e Mon Sep 17 00:00:00 2001 From: Pavel Tisnovsky Date: Wed, 6 Aug 2025 08:57:41 +0200 Subject: [PATCH 2/5] Fixed imports --- src/client.py | 2 +- src/utils/common.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/src/client.py b/src/client.py index 93415008..04ddef41 100644 --- a/src/client.py +++ b/src/client.py @@ -4,7 +4,7 @@ from typing import Optional -from llama_stack.distribution.library_client import ( +from llama_stack import ( AsyncLlamaStackAsLibraryClient, # type: ignore LlamaStackAsLibraryClient, # type: ignore ) diff --git a/src/utils/common.py b/src/utils/common.py index 47ec67cb..e4a35a57 100644 --- a/src/utils/common.py +++ b/src/utils/common.py @@ -6,7 +6,7 @@ from logging import Logger from llama_stack_client import LlamaStackClient, AsyncLlamaStackClient -from llama_stack.distribution.library_client import ( +from llama_stack import ( AsyncLlamaStackAsLibraryClient, ) From 1767d6b0a5f808fc98697b46ff5604a541bf2de0 Mon Sep 17 00:00:00 2001 From: Pavel Tisnovsky Date: Wed, 6 Aug 2025 11:15:02 +0200 Subject: [PATCH 3/5] Tests needs to use 0.2.17 too --- test.containerfile | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/test.containerfile b/test.containerfile index b6b198a1..6abc4d6a 100644 --- a/test.containerfile +++ b/test.containerfile @@ -17,7 +17,7 @@ RUN curl -LsSf https://astral.sh/uv/install.sh | sh RUN uv -h RUN uv venv && \ - uv pip install llama-stack==0.2.16 \ + uv pip install llama-stack==0.2.17 \ fastapi \ opentelemetry-sdk \ opentelemetry-exporter-otlp \ @@ -36,4 +36,4 @@ RUN uv venv && \ peft \ trl -CMD ["uv", "run", "llama", "stack", "run", "run.yaml"] \ No newline at end of file +CMD ["uv", "run", "llama", "stack", "run", "run.yaml"] From 2a0e4170e6b6332f8c91bcd4e71715631774759e Mon Sep 17 00:00:00 2001 From: Radovan Fuchs Date: Wed, 6 Aug 2025 11:42:39 +0200 Subject: [PATCH 4/5] use llama-stack from pyproject --- test.containerfile | 8 +++----- 1 file changed, 3 insertions(+), 5 deletions(-) diff --git a/test.containerfile b/test.containerfile index 6abc4d6a..988ac6ca 100644 --- a/test.containerfile +++ b/test.containerfile @@ -7,7 +7,7 @@ ENV PATH="$PATH:/root/.local/bin" WORKDIR ${APP_ROOT} COPY run.yaml ./ - +COPY pyproject.toml uv.lock ./ RUN microdnf install -y --nodocs --setopt=keepcache=0 --setopt=tsflags=nodocs \ python3.12 python3.12-devel python3.12-pip git tar @@ -16,15 +16,13 @@ RUN curl -LsSf https://astral.sh/uv/install.sh | sh RUN uv -h -RUN uv venv && \ - uv pip install llama-stack==0.2.17 \ - fastapi \ +RUN uv sync --no-install-project --no-dev && \ + uv pip install \ opentelemetry-sdk \ opentelemetry-exporter-otlp \ opentelemetry-instrumentation \ aiosqlite \ litellm \ - uvicorn \ blobfile \ datasets \ sqlalchemy \ From a90e0c5a09cd2a6064de032f6568a90ecaa7bf00 Mon Sep 17 00:00:00 2001 From: Radovan Fuchs Date: Wed, 6 Aug 2025 12:27:27 +0200 Subject: [PATCH 5/5] revert --- test.containerfile | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/test.containerfile b/test.containerfile index 988ac6ca..cd6f5785 100644 --- a/test.containerfile +++ b/test.containerfile @@ -7,7 +7,6 @@ ENV PATH="$PATH:/root/.local/bin" WORKDIR ${APP_ROOT} COPY run.yaml ./ -COPY pyproject.toml uv.lock ./ RUN microdnf install -y --nodocs --setopt=keepcache=0 --setopt=tsflags=nodocs \ python3.12 python3.12-devel python3.12-pip git tar @@ -16,13 +15,15 @@ RUN curl -LsSf https://astral.sh/uv/install.sh | sh RUN uv -h -RUN uv sync --no-install-project --no-dev && \ - uv pip install \ +RUN uv venv && \ + uv pip install llama-stack==0.2.17 \ + fastapi \ opentelemetry-sdk \ opentelemetry-exporter-otlp \ opentelemetry-instrumentation \ aiosqlite \ litellm \ + uvicorn \ blobfile \ datasets \ sqlalchemy \