Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
17 commits
Select commit Hold shift + click to select a range
c94275d
fix: avoid newer type syntax
stainless-app[bot] Aug 27, 2025
c101839
chore(internal): run tests in an isolated environment
stainless-app[bot] Sep 18, 2025
402a316
chore(internal): update pyright exclude list
stainless-app[bot] Sep 18, 2025
1956d9a
chore(internal): minor formatting change
stainless-app[bot] Aug 27, 2025
54fd160
chore(internal): add Sequence related utils
stainless-app[bot] Aug 30, 2025
2c0cdd1
feat(types): replace List[str] with SequenceNotStr in params
stainless-app[bot] Sep 3, 2025
5c7b05b
feat: improve future compat with pydantic v3
stainless-app[bot] Sep 4, 2025
a27daa1
chore(internal): move mypy configurations to `pyproject.toml` file
stainless-app[bot] Sep 5, 2025
fc50cbe
fix(client): fix circular dependencies and offset pagination
stainless-app[bot] Sep 5, 2025
1cf74fe
chore(internal): codegen related update
stainless-app[bot] Sep 6, 2025
b48f1f4
chore(internal): update pydantic dependency
stainless-app[bot] Sep 17, 2025
c241e7e
chore(types): change optional parameter type from NotGiven to Omit
stainless-app[bot] Sep 19, 2025
6f269ff
chore: do not install brew dependencies in ./scripts/bootstrap by def…
stainless-app[bot] Sep 20, 2025
a2f4544
chore(deps): move deprecated `dev-dependencies` in `pyproject.toml` t…
stainless-app[bot] Sep 24, 2025
e9873f9
feat(api): manual updates
stainless-app[bot] Sep 25, 2025
ccc9dd2
fix(project): avoid duplicate dev dep sections
RobertCraigie Sep 26, 2025
d8178cf
release: 0.2.23-alpha.1
stainless-app[bot] Sep 26, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .release-please-manifest.json
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
{
".": "0.2.19-alpha.1"
".": "0.2.23-alpha.1"
}
4 changes: 2 additions & 2 deletions .stats.yml
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
configured_endpoints: 107
configured_endpoints: 111
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-f252873ea1e1f38fd207331ef2621c511154d5be3f4076e59cc15754fc58eee4.yml
openapi_spec_hash: 10cbb4337a06a9fdd7d08612dd6044c3
config_hash: 40b8d777e1eb8b6ab05759b663edd2fb
config_hash: 0358112cc0f3d880b4d55debdbe1cfa3
40 changes: 40 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,45 @@
# Changelog

## 0.2.23-alpha.1 (2025-09-26)

Full Changelog: [v0.2.19-alpha.1...v0.2.23-alpha.1](https://github.com/llamastack/llama-stack-client-python/compare/v0.2.19-alpha.1...v0.2.23-alpha.1)

### Features

* **api:** manual updates ([e9873f9](https://github.com/llamastack/llama-stack-client-python/commit/e9873f92a90262e76e67b50a054c3632dddab572))
* improve future compat with pydantic v3 ([5c7b05b](https://github.com/llamastack/llama-stack-client-python/commit/5c7b05b5bcb922d82b57838f31a734588f31855a))
* **types:** replace List[str] with SequenceNotStr in params ([2c0cdd1](https://github.com/llamastack/llama-stack-client-python/commit/2c0cdd14a84245001b2f46d220950b7828e4675d))


### Bug Fixes

* avoid newer type syntax ([c94275d](https://github.com/llamastack/llama-stack-client-python/commit/c94275d8aea13deafb0821e535774f48f0063a37))
* **client:** fix circular dependencies and offset pagination ([fc50cbe](https://github.com/llamastack/llama-stack-client-python/commit/fc50cbe63b0483b1ac786e09134ae8917d0a40b7))
* **project:** avoid duplicate dev dep sections ([ccc9dd2](https://github.com/llamastack/llama-stack-client-python/commit/ccc9dd2d704424b0b1f6a3f1d0da09385bfda5cf))
* **tool:** revert the ToolDefParam stuff because we reverted the bad commit ([e2d543d](https://github.com/llamastack/llama-stack-client-python/commit/e2d543dfe2cdab2e1d74e3fa2bd2334af040b56d))


### Chores

* **deps:** move deprecated `dev-dependencies` in `pyproject.toml` to dev group ([a2f4544](https://github.com/llamastack/llama-stack-client-python/commit/a2f45441edf67842c2db37765ee773394cd2dda3))
* do not install brew dependencies in ./scripts/bootstrap by default ([6f269ff](https://github.com/llamastack/llama-stack-client-python/commit/6f269ff5dec08256d4e2e144b85cd67c8f860bb2))
* **internal:** add Sequence related utils ([54fd160](https://github.com/llamastack/llama-stack-client-python/commit/54fd1607f0dc4506a329b8a8f3c747cc6a413789))
* **internal:** codegen related update ([1cf74fe](https://github.com/llamastack/llama-stack-client-python/commit/1cf74fec41180b9357ad91cf6f67b950c98db150))
* **internal:** minor formatting change ([1956d9a](https://github.com/llamastack/llama-stack-client-python/commit/1956d9aaf2aec0ab8bee590dc8df3d049689acd9))
* **internal:** move mypy configurations to `pyproject.toml` file ([a27daa1](https://github.com/llamastack/llama-stack-client-python/commit/a27daa1627b674eb431175629f9be54e9fc8a39e))
* **internal:** run tests in an isolated environment ([c101839](https://github.com/llamastack/llama-stack-client-python/commit/c1018390ca2754d8d6152bc3a350e95ce1138ac9))
* **internal:** update pydantic dependency ([b48f1f4](https://github.com/llamastack/llama-stack-client-python/commit/b48f1f4d730751de94029733e9466edd08589464))
* **internal:** update pyright exclude list ([402a316](https://github.com/llamastack/llama-stack-client-python/commit/402a3160399d676cdcd214ea8a776fd2ce7c2db3))
* **types:** change optional parameter type from NotGiven to Omit ([c241e7e](https://github.com/llamastack/llama-stack-client-python/commit/c241e7e5f4986b304bf9003eaf3224c7b141a500))


### Build System

* Bump version to 0.2.19 ([034b3a9](https://github.com/llamastack/llama-stack-client-python/commit/034b3a9debdb85be0cbe42cb8711e37629d3376c))
* Bump version to 0.2.20 ([b178953](https://github.com/llamastack/llama-stack-client-python/commit/b178953c51e1bfe5075fe942eacab90e0caad180))
* Bump version to 0.2.21 ([bb25f4a](https://github.com/llamastack/llama-stack-client-python/commit/bb25f4aab73df2f336a3f3ff184476367a1b66b4))
* Bump version to 0.2.22 ([ba825eb](https://github.com/llamastack/llama-stack-client-python/commit/ba825eb25975fa7ae8589365871e59d1b02168d8))

## 0.2.19-alpha.1 (2025-08-26)

Full Changelog: [v0.2.18-alpha.3...v0.2.19-alpha.1](https://github.com/llamastack/llama-stack-client-python/compare/v0.2.18-alpha.3...v0.2.19-alpha.1)
Expand Down
31 changes: 26 additions & 5 deletions api.md
Original file line number Diff line number Diff line change
Expand Up @@ -81,14 +81,20 @@ Methods:
Types:

```python
from llama_stack_client.types import ResponseObject, ResponseObjectStream, ResponseListResponse
from llama_stack_client.types import (
ResponseObject,
ResponseObjectStream,
ResponseListResponse,
ResponseDeleteResponse,
)
```

Methods:

- <code title="post /v1/openai/v1/responses">client.responses.<a href="./src/llama_stack_client/resources/responses/responses.py">create</a>(\*\*<a href="src/llama_stack_client/types/response_create_params.py">params</a>) -> <a href="./src/llama_stack_client/types/response_object.py">ResponseObject</a></code>
- <code title="get /v1/openai/v1/responses/{response_id}">client.responses.<a href="./src/llama_stack_client/resources/responses/responses.py">retrieve</a>(response_id) -> <a href="./src/llama_stack_client/types/response_object.py">ResponseObject</a></code>
- <code title="get /v1/openai/v1/responses">client.responses.<a href="./src/llama_stack_client/resources/responses/responses.py">list</a>(\*\*<a href="src/llama_stack_client/types/response_list_params.py">params</a>) -> <a href="./src/llama_stack_client/types/response_list_response.py">SyncOpenAICursorPage[ResponseListResponse]</a></code>
- <code title="delete /v1/openai/v1/responses/{response_id}">client.responses.<a href="./src/llama_stack_client/resources/responses/responses.py">delete</a>(response_id) -> <a href="./src/llama_stack_client/types/response_delete_response.py">ResponseDeleteResponse</a></code>

## InputItems

Expand Down Expand Up @@ -242,6 +248,7 @@ from llama_stack_client.types import (
EmbeddingsResponse,
TokenLogProbs,
InferenceBatchChatCompletionResponse,
InferenceRerankResponse,
)
```

Expand All @@ -252,6 +259,7 @@ Methods:
- <code title="post /v1/inference/chat-completion">client.inference.<a href="./src/llama_stack_client/resources/inference.py">chat_completion</a>(\*\*<a href="src/llama_stack_client/types/inference_chat_completion_params.py">params</a>) -> <a href="./src/llama_stack_client/types/shared/chat_completion_response.py">ChatCompletionResponse</a></code>
- <code title="post /v1/inference/completion">client.inference.<a href="./src/llama_stack_client/resources/inference.py">completion</a>(\*\*<a href="src/llama_stack_client/types/inference_completion_params.py">params</a>) -> <a href="./src/llama_stack_client/types/completion_response.py">CompletionResponse</a></code>
- <code title="post /v1/inference/embeddings">client.inference.<a href="./src/llama_stack_client/resources/inference.py">embeddings</a>(\*\*<a href="src/llama_stack_client/types/inference_embeddings_params.py">params</a>) -> <a href="./src/llama_stack_client/types/embeddings_response.py">EmbeddingsResponse</a></code>
- <code title="post /v1/inference/rerank">client.inference.<a href="./src/llama_stack_client/resources/inference.py">rerank</a>(\*\*<a href="src/llama_stack_client/types/inference_rerank_params.py">params</a>) -> <a href="./src/llama_stack_client/types/inference_rerank_response.py">InferenceRerankResponse</a></code>

# Embeddings

Expand Down Expand Up @@ -389,10 +397,22 @@ from llama_stack_client.types import ListModelsResponse, Model, ModelListRespons

Methods:

- <code title="get /v1/models/{model_id}">client.models.<a href="./src/llama_stack_client/resources/models.py">retrieve</a>(model_id) -> <a href="./src/llama_stack_client/types/model.py">Model</a></code>
- <code title="get /v1/models">client.models.<a href="./src/llama_stack_client/resources/models.py">list</a>() -> <a href="./src/llama_stack_client/types/model_list_response.py">ModelListResponse</a></code>
- <code title="post /v1/models">client.models.<a href="./src/llama_stack_client/resources/models.py">register</a>(\*\*<a href="src/llama_stack_client/types/model_register_params.py">params</a>) -> <a href="./src/llama_stack_client/types/model.py">Model</a></code>
- <code title="delete /v1/models/{model_id}">client.models.<a href="./src/llama_stack_client/resources/models.py">unregister</a>(model_id) -> None</code>
- <code title="get /v1/models/{model_id}">client.models.<a href="./src/llama_stack_client/resources/models/models.py">retrieve</a>(model_id) -> <a href="./src/llama_stack_client/types/model.py">Model</a></code>
- <code title="get /v1/models">client.models.<a href="./src/llama_stack_client/resources/models/models.py">list</a>() -> <a href="./src/llama_stack_client/types/model_list_response.py">ModelListResponse</a></code>
- <code title="post /v1/models">client.models.<a href="./src/llama_stack_client/resources/models/models.py">register</a>(\*\*<a href="src/llama_stack_client/types/model_register_params.py">params</a>) -> <a href="./src/llama_stack_client/types/model.py">Model</a></code>
- <code title="delete /v1/models/{model_id}">client.models.<a href="./src/llama_stack_client/resources/models/models.py">unregister</a>(model_id) -> None</code>

## OpenAI

Types:

```python
from llama_stack_client.types.models import OpenAIListResponse
```

Methods:

- <code title="get /v1/openai/v1/models">client.models.openai.<a href="./src/llama_stack_client/resources/models/openai.py">list</a>() -> <a href="./src/llama_stack_client/types/models/openai_list_response.py">OpenAIListResponse</a></code>

# PostTraining

Expand Down Expand Up @@ -487,6 +507,7 @@ Methods:

- <code title="get /v1/shields/{identifier}">client.shields.<a href="./src/llama_stack_client/resources/shields.py">retrieve</a>(identifier) -> <a href="./src/llama_stack_client/types/shield.py">Shield</a></code>
- <code title="get /v1/shields">client.shields.<a href="./src/llama_stack_client/resources/shields.py">list</a>() -> <a href="./src/llama_stack_client/types/shield_list_response.py">ShieldListResponse</a></code>
- <code title="delete /v1/shields/{identifier}">client.shields.<a href="./src/llama_stack_client/resources/shields.py">delete</a>(identifier) -> None</code>
- <code title="post /v1/shields">client.shields.<a href="./src/llama_stack_client/resources/shields.py">register</a>(\*\*<a href="src/llama_stack_client/types/shield_register_params.py">params</a>) -> <a href="./src/llama_stack_client/types/shield.py">Shield</a></code>

# SyntheticDataGeneration
Expand Down
50 changes: 0 additions & 50 deletions mypy.ini

This file was deleted.

88 changes: 66 additions & 22 deletions pyproject.toml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
[project]
name = "llama_stack_client"
version = "0.2.22"
version = "0.2.23-alpha.1"
description = "The official Python library for the llama-stack-client API"
dynamic = ["readme"]
license = "MIT"
Expand Down Expand Up @@ -46,45 +46,36 @@ aiohttp = ["aiohttp", "httpx_aiohttp>=0.1.8"]
[tool.uv]
managed = true
required-version = ">=0.5.0"
conflicts = [
[
{ group = "pydantic-v1" },
{ group = "pydantic-v2" },
],
]

[dependency-groups]
# version pins are in uv.lock
dev-dependencies = [
dev = [
"pyright==1.1.399",
"mypy",
"respx",
"pytest",
"pytest>=7.1.1",
"pytest-asyncio",
"ruff",
"time-machine",
"dirty-equals>=0.6.0",
"importlib-metadata>=6.7.0",
"rich>=13.7.1",
"nest_asyncio==1.6.0",
"pytest-xdist>=3.6.1",
"pre-commit",
"black",
]
conflicts = [
[
{ group = "pydantic-v1" },
{ group = "pydantic-v2" },
],
]

[dependency-groups]
pydantic-v1 = [
"pydantic>=1.9.0,<2",
]
pydantic-v2 = [
"pydantic>=2,<3",
]
dev = [
"pytest>=7.1.1",
"pytest-asyncio",
"pre-commit",
"black",
"ruff",
"mypy",
"respx",
"dirty-equals"
]


[build-system]
Expand Down Expand Up @@ -135,6 +126,59 @@ filterwarnings = [
"error"
]

[tool.mypy]
pretty = true
show_error_codes = true

# Exclude _files.py because mypy isn't smart enough to apply
# the correct type narrowing and as this is an internal module
# it's fine to just use Pyright.
#
# We also exclude our `tests` as mypy doesn't always infer
# types correctly and Pyright will still catch any type errors.
exclude = ['src/llama_stack_client/_files.py', '_dev/.*.py', 'tests/.*']

strict_equality = true
implicit_reexport = true
check_untyped_defs = true
no_implicit_optional = true

warn_return_any = true
warn_unreachable = true
warn_unused_configs = true

# Turn these options off as it could cause conflicts
# with the Pyright options.
warn_unused_ignores = false
warn_redundant_casts = false

disallow_any_generics = true
disallow_untyped_defs = true
disallow_untyped_calls = true
disallow_subclassing_any = true
disallow_incomplete_defs = true
disallow_untyped_decorators = true
cache_fine_grained = true

# By default, mypy reports an error if you assign a value to the result
# of a function call that doesn't return anything. We do this in our test
# cases:
# ```
# result = ...
# assert result is None
# ```
# Changing this codegen to make mypy happy would increase complexity
# and would not be worth it.
disable_error_code = "func-returns-value,overload-cannot-match"

# https://github.com/python/mypy/issues/12162

[[tool.mypy.overrides]]
module = "black.files.*"
ignore_errors = true
ignore_missing_imports = true


[tool.ruff]
line-length = 120
output-format = "grouped"
Expand Down
Loading
Loading