Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions .stats.yml
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
configured_endpoints: 104
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-97b91eca4a3ff251edc02636b1a638866675d6c1abd46cd9fc18bc50a1de9656.yml
openapi_spec_hash: 7302f1aa50090e3de78e34c184371267
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-35c6569e5e9fcc85084c9728eb7fc7c5908297fcc77043d621d25de3c850a990.yml
openapi_spec_hash: 0f95bbeee16f3205d36ec34cfa62c711
config_hash: a3829dbdaa491194d01f399784d532cd
10 changes: 5 additions & 5 deletions src/llama_stack_client/resources/moderations.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@
import httpx

from ..types import moderation_create_params
from .._types import Body, Query, Headers, NotGiven, SequenceNotStr, not_given
from .._types import Body, Omit, Query, Headers, NotGiven, SequenceNotStr, omit, not_given
from .._utils import maybe_transform, async_maybe_transform
from .._compat import cached_property
from .._resource import SyncAPIResource, AsyncAPIResource
Expand Down Expand Up @@ -53,7 +53,7 @@ def create(
self,
*,
input: Union[str, SequenceNotStr[str]],
model: str,
model: str | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
Expand All @@ -70,7 +70,7 @@ def create(
input: Input (or inputs) to classify. Can be a single string, an array of strings, or
an array of multi-modal input objects similar to other models.

model: The content moderation model you would like to use.
model: (Optional) The content moderation model you would like to use.

extra_headers: Send extra headers

Expand Down Expand Up @@ -120,7 +120,7 @@ async def create(
self,
*,
input: Union[str, SequenceNotStr[str]],
model: str,
model: str | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
Expand All @@ -137,7 +137,7 @@ async def create(
input: Input (or inputs) to classify. Can be a single string, an array of strings, or
an array of multi-modal input objects similar to other models.

model: The content moderation model you would like to use.
model: (Optional) The content moderation model you would like to use.

extra_headers: Send extra headers

Expand Down
24 changes: 12 additions & 12 deletions src/llama_stack_client/resources/tool_runtime/rag_tool.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@ def insert(
*,
chunk_size_in_tokens: int,
documents: Iterable[Document],
vector_db_id: str,
vector_store_id: str,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
Expand All @@ -73,7 +73,7 @@ def insert(

documents: List of documents to index in the RAG system

vector_db_id: ID of the vector database to store the document embeddings
vector_store_id: ID of the vector database to store the document embeddings

extra_headers: Send extra headers

Expand All @@ -90,7 +90,7 @@ def insert(
{
"chunk_size_in_tokens": chunk_size_in_tokens,
"documents": documents,
"vector_db_id": vector_db_id,
"vector_store_id": vector_store_id,
},
rag_tool_insert_params.RagToolInsertParams,
),
Expand All @@ -104,7 +104,7 @@ def query(
self,
*,
content: InterleavedContent,
vector_db_ids: SequenceNotStr[str],
vector_store_ids: SequenceNotStr[str],
query_config: QueryConfig | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
Expand All @@ -119,7 +119,7 @@ def query(
Args:
content: The query content to search for in the indexed documents

vector_db_ids: List of vector database IDs to search within
vector_store_ids: List of vector database IDs to search within

query_config: (Optional) Configuration parameters for the query operation

Expand All @@ -136,7 +136,7 @@ def query(
body=maybe_transform(
{
"content": content,
"vector_db_ids": vector_db_ids,
"vector_store_ids": vector_store_ids,
"query_config": query_config,
},
rag_tool_query_params.RagToolQueryParams,
Expand Down Expand Up @@ -173,7 +173,7 @@ async def insert(
*,
chunk_size_in_tokens: int,
documents: Iterable[Document],
vector_db_id: str,
vector_store_id: str,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
Expand All @@ -189,7 +189,7 @@ async def insert(

documents: List of documents to index in the RAG system

vector_db_id: ID of the vector database to store the document embeddings
vector_store_id: ID of the vector database to store the document embeddings

extra_headers: Send extra headers

Expand All @@ -206,7 +206,7 @@ async def insert(
{
"chunk_size_in_tokens": chunk_size_in_tokens,
"documents": documents,
"vector_db_id": vector_db_id,
"vector_store_id": vector_store_id,
},
rag_tool_insert_params.RagToolInsertParams,
),
Expand All @@ -220,7 +220,7 @@ async def query(
self,
*,
content: InterleavedContent,
vector_db_ids: SequenceNotStr[str],
vector_store_ids: SequenceNotStr[str],
query_config: QueryConfig | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
Expand All @@ -235,7 +235,7 @@ async def query(
Args:
content: The query content to search for in the indexed documents

vector_db_ids: List of vector database IDs to search within
vector_store_ids: List of vector database IDs to search within

query_config: (Optional) Configuration parameters for the query operation

Expand All @@ -252,7 +252,7 @@ async def query(
body=await async_maybe_transform(
{
"content": content,
"vector_db_ids": vector_db_ids,
"vector_store_ids": vector_store_ids,
"query_config": query_config,
},
rag_tool_query_params.RagToolQueryParams,
Expand Down
24 changes: 12 additions & 12 deletions src/llama_stack_client/resources/vector_io.py
Original file line number Diff line number Diff line change
Expand Up @@ -54,7 +54,7 @@ def insert(
self,
*,
chunks: Iterable[vector_io_insert_params.Chunk],
vector_db_id: str,
vector_store_id: str,
ttl_seconds: int | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
Expand All @@ -74,7 +74,7 @@ def insert(
configure how Llama Stack formats the chunk during generation. If `embedding` is
not provided, it will be computed later.

vector_db_id: The identifier of the vector database to insert the chunks into.
vector_store_id: The identifier of the vector database to insert the chunks into.

ttl_seconds: The time to live of the chunks.

Expand All @@ -92,7 +92,7 @@ def insert(
body=maybe_transform(
{
"chunks": chunks,
"vector_db_id": vector_db_id,
"vector_store_id": vector_store_id,
"ttl_seconds": ttl_seconds,
},
vector_io_insert_params.VectorIoInsertParams,
Expand All @@ -107,7 +107,7 @@ def query(
self,
*,
query: InterleavedContent,
vector_db_id: str,
vector_store_id: str,
params: Dict[str, Union[bool, float, str, Iterable[object], object, None]] | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
Expand All @@ -122,7 +122,7 @@ def query(
Args:
query: The query to search for.

vector_db_id: The identifier of the vector database to query.
vector_store_id: The identifier of the vector database to query.

params: The parameters of the query.

Expand All @@ -139,7 +139,7 @@ def query(
body=maybe_transform(
{
"query": query,
"vector_db_id": vector_db_id,
"vector_store_id": vector_store_id,
"params": params,
},
vector_io_query_params.VectorIoQueryParams,
Expand Down Expand Up @@ -175,7 +175,7 @@ async def insert(
self,
*,
chunks: Iterable[vector_io_insert_params.Chunk],
vector_db_id: str,
vector_store_id: str,
ttl_seconds: int | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
Expand All @@ -195,7 +195,7 @@ async def insert(
configure how Llama Stack formats the chunk during generation. If `embedding` is
not provided, it will be computed later.

vector_db_id: The identifier of the vector database to insert the chunks into.
vector_store_id: The identifier of the vector database to insert the chunks into.

ttl_seconds: The time to live of the chunks.

Expand All @@ -213,7 +213,7 @@ async def insert(
body=await async_maybe_transform(
{
"chunks": chunks,
"vector_db_id": vector_db_id,
"vector_store_id": vector_store_id,
"ttl_seconds": ttl_seconds,
},
vector_io_insert_params.VectorIoInsertParams,
Expand All @@ -228,7 +228,7 @@ async def query(
self,
*,
query: InterleavedContent,
vector_db_id: str,
vector_store_id: str,
params: Dict[str, Union[bool, float, str, Iterable[object], object, None]] | Omit = omit,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
Expand All @@ -243,7 +243,7 @@ async def query(
Args:
query: The query to search for.

vector_db_id: The identifier of the vector database to query.
vector_store_id: The identifier of the vector database to query.

params: The parameters of the query.

Expand All @@ -260,7 +260,7 @@ async def query(
body=await async_maybe_transform(
{
"query": query,
"vector_db_id": vector_db_id,
"vector_store_id": vector_store_id,
"params": params,
},
vector_io_query_params.VectorIoQueryParams,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@ class MemoryRetrievalStep(BaseModel):
turn_id: str
"""The ID of the turn."""

vector_db_ids: str
vector_store_ids: str
"""The IDs of the vector databases to retrieve context from."""

completed_at: Optional[datetime] = None
Expand Down
4 changes: 2 additions & 2 deletions src/llama_stack_client/types/moderation_create_params.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,5 +24,5 @@ class ModerationCreateParams(TypedDict, total=False):
objects similar to other models.
"""

model: Required[str]
"""The content moderation model you would like to use."""
model: str
"""(Optional) The content moderation model you would like to use."""
Loading
Loading