From 742638a2d60f24e533050ff0013e91e576c9baab Mon Sep 17 00:00:00 2001 From: Pavel Tisnovsky Date: Thu, 2 Oct 2025 12:41:48 +0200 Subject: [PATCH] LCORE-695: updated doc --- README.md | 2 +- docs/deployment_guide.md | 6 +++--- docs/getting_started.md | 2 +- docs/openapi.json | 7 ++++--- docs/openapi.md | 2 +- docs/output.md | 2 +- examples/pyproject.llamastack.toml | 2 +- src/models/responses.py | 4 ++-- 8 files changed, 14 insertions(+), 13 deletions(-) diff --git a/README.md b/README.md index 505190f3..d4d7ab35 100644 --- a/README.md +++ b/README.md @@ -245,7 +245,7 @@ version = "0.1.0" description = "Llama Stack runner" authors = [] dependencies = [ - "llama-stack==0.2.21", + "llama-stack==0.2.22", "fastapi>=0.115.12", "opentelemetry-sdk>=1.34.0", "opentelemetry-exporter-otlp>=1.34.0", diff --git a/docs/deployment_guide.md b/docs/deployment_guide.md index 296718d4..a619d8a3 100644 --- a/docs/deployment_guide.md +++ b/docs/deployment_guide.md @@ -390,7 +390,7 @@ cp examples/run.yaml /tmp/llama-stack-server The output should be in this form: ```json { - "version": "0.2.21" + "version": "0.2.22" } ``` @@ -676,7 +676,7 @@ a4982f43195537b9eb1cec510fe6655f245d6d4b7236a4759808115d5d719972 description = "Default template for PDM package" authors = [] dependencies = [ - "llama-stack==0.2.18", + "llama-stack==0.2.22", "fastapi>=0.115.12", "opentelemetry-sdk>=1.34.0", "opentelemetry-exporter-otlp>=1.34.0", @@ -1071,7 +1071,7 @@ models: The output should be in this form: ```json { - "version": "0.2.18" + "version": "0.2.22" } ``` diff --git a/docs/getting_started.md b/docs/getting_started.md index 5752c715..d75596f3 100644 --- a/docs/getting_started.md +++ b/docs/getting_started.md @@ -24,7 +24,7 @@ It is possible to run Lightspeed Core Stack service with Llama Stack "embedded" 1. Add and install all required dependencies ```bash uv add \ - "llama-stack==0.2.21" \ + "llama-stack==0.2.22" \ "fastapi>=0.115.12" \ "opentelemetry-sdk>=1.34.0" \ "opentelemetry-exporter-otlp>=1.34.0" \ diff --git a/docs/openapi.json b/docs/openapi.json index 608b1c1f..20a5a5bb 100644 --- a/docs/openapi.json +++ b/docs/openapi.json @@ -2142,7 +2142,8 @@ "examples": [ "0.2.1", "0.2.2", - "0.2.18" + "0.2.18", + "0.2.22" ] } }, @@ -2153,7 +2154,7 @@ "llama_stack_version" ], "title": "InfoResponse", - "description": "Model representing a response to an info request.\n\nAttributes:\n name: Service name.\n service_version: Service version.\n llama_stack_version: Llama Stack version.\n\nExample:\n ```python\n info_response = InfoResponse(\n name=\"Lightspeed Stack\",\n service_version=\"1.0.0\",\n llama_stack_version=\"0.2.21\",\n )\n ```", + "description": "Model representing a response to an info request.\n\nAttributes:\n name: Service name.\n service_version: Service version.\n llama_stack_version: Llama Stack version.\n\nExample:\n ```python\n info_response = InfoResponse(\n name=\"Lightspeed Stack\",\n service_version=\"1.0.0\",\n llama_stack_version=\"0.2.22\",\n )\n ```", "examples": [ { "llama_stack_version": "1.0.0", @@ -3159,4 +3160,4 @@ } } } -} \ No newline at end of file +} diff --git a/docs/openapi.md b/docs/openapi.md index b70a7f94..f2aaf8f1 100644 --- a/docs/openapi.md +++ b/docs/openapi.md @@ -1039,7 +1039,7 @@ Example: info_response = InfoResponse( name="Lightspeed Stack", service_version="1.0.0", - llama_stack_version="0.2.21", + llama_stack_version="0.2.22", ) ``` diff --git a/docs/output.md b/docs/output.md index b70a7f94..f2aaf8f1 100644 --- a/docs/output.md +++ b/docs/output.md @@ -1039,7 +1039,7 @@ Example: info_response = InfoResponse( name="Lightspeed Stack", service_version="1.0.0", - llama_stack_version="0.2.21", + llama_stack_version="0.2.22", ) ``` diff --git a/examples/pyproject.llamastack.toml b/examples/pyproject.llamastack.toml index fd4ae2a4..9f873ee0 100644 --- a/examples/pyproject.llamastack.toml +++ b/examples/pyproject.llamastack.toml @@ -4,7 +4,7 @@ version = "0.1.0" description = "Default template for PDM package" authors = [] dependencies = [ - "llama-stack==0.2.21", + "llama-stack==0.2.22", "fastapi>=0.115.12", "opentelemetry-sdk>=1.34.0", "opentelemetry-exporter-otlp>=1.34.0", diff --git a/src/models/responses.py b/src/models/responses.py index 7674b884..c2931d66 100644 --- a/src/models/responses.py +++ b/src/models/responses.py @@ -164,7 +164,7 @@ class InfoResponse(BaseModel): info_response = InfoResponse( name="Lightspeed Stack", service_version="1.0.0", - llama_stack_version="0.2.21", + llama_stack_version="0.2.22", ) ``` """ @@ -181,7 +181,7 @@ class InfoResponse(BaseModel): llama_stack_version: str = Field( description="Llama Stack version", - examples=["0.2.1", "0.2.2", "0.2.18"], + examples=["0.2.1", "0.2.2", "0.2.18", "0.2.21", "0.2.22"], ) # provides examples for /docs endpoint