Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions api.md
Original file line number Diff line number Diff line change
Expand Up @@ -69,12 +69,12 @@ Methods:
Types:

```python
from openlayer.types.inference_pipelines import RowStreamResponse
from openlayer.types.inference_pipelines import RowUpdateResponse
```

Methods:

- <code title="put /inference-pipelines/{inferencePipelineId}/rows">client.inference_pipelines.rows.<a href="./src/openlayer/resources/inference_pipelines/rows.py">stream</a>(inference_pipeline_id, \*\*<a href="src/openlayer/types/inference_pipelines/row_stream_params.py">params</a>) -> <a href="./src/openlayer/types/inference_pipelines/row_stream_response.py">RowStreamResponse</a></code>
- <code title="put /inference-pipelines/{inferencePipelineId}/rows">client.inference_pipelines.rows.<a href="./src/openlayer/resources/inference_pipelines/rows.py">update</a>(inference_pipeline_id, \*\*<a href="src/openlayer/types/inference_pipelines/row_update_params.py">params</a>) -> <a href="./src/openlayer/types/inference_pipelines/row_update_response.py">RowUpdateResponse</a></code>

## TestResults

Expand Down
44 changes: 22 additions & 22 deletions src/openlayer/resources/inference_pipelines/rows.py
Original file line number Diff line number Diff line change
Expand Up @@ -20,8 +20,8 @@
async_to_streamed_response_wrapper,
)
from ..._base_client import make_request_options
from ...types.inference_pipelines import row_stream_params
from ...types.inference_pipelines.row_stream_response import RowStreamResponse
from ...types.inference_pipelines import row_update_params
from ...types.inference_pipelines.row_update_response import RowUpdateResponse

__all__ = ["RowsResource", "AsyncRowsResource"]

Expand All @@ -35,20 +35,20 @@ def with_raw_response(self) -> RowsResourceWithRawResponse:
def with_streaming_response(self) -> RowsResourceWithStreamingResponse:
return RowsResourceWithStreamingResponse(self)

def stream(
def update(
self,
inference_pipeline_id: str,
*,
inference_id: str,
row: object,
config: Optional[row_stream_params.Config] | NotGiven = NOT_GIVEN,
config: Optional[row_update_params.Config] | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
extra_query: Query | None = None,
extra_body: Body | None = None,
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
) -> RowStreamResponse:
) -> RowUpdateResponse:
"""
Update an inference data point in an inference pipeline.

Expand All @@ -74,16 +74,16 @@ def stream(
"row": row,
"config": config,
},
row_stream_params.RowStreamParams,
row_update_params.RowUpdateParams,
),
options=make_request_options(
extra_headers=extra_headers,
extra_query=extra_query,
extra_body=extra_body,
timeout=timeout,
query=maybe_transform({"inference_id": inference_id}, row_stream_params.RowStreamParams),
query=maybe_transform({"inference_id": inference_id}, row_update_params.RowUpdateParams),
),
cast_to=RowStreamResponse,
cast_to=RowUpdateResponse,
)


Expand All @@ -96,20 +96,20 @@ def with_raw_response(self) -> AsyncRowsResourceWithRawResponse:
def with_streaming_response(self) -> AsyncRowsResourceWithStreamingResponse:
return AsyncRowsResourceWithStreamingResponse(self)

async def stream(
async def update(
self,
inference_pipeline_id: str,
*,
inference_id: str,
row: object,
config: Optional[row_stream_params.Config] | NotGiven = NOT_GIVEN,
config: Optional[row_update_params.Config] | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
extra_query: Query | None = None,
extra_body: Body | None = None,
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
) -> RowStreamResponse:
) -> RowUpdateResponse:
"""
Update an inference data point in an inference pipeline.

Expand All @@ -135,50 +135,50 @@ async def stream(
"row": row,
"config": config,
},
row_stream_params.RowStreamParams,
row_update_params.RowUpdateParams,
),
options=make_request_options(
extra_headers=extra_headers,
extra_query=extra_query,
extra_body=extra_body,
timeout=timeout,
query=await async_maybe_transform({"inference_id": inference_id}, row_stream_params.RowStreamParams),
query=await async_maybe_transform({"inference_id": inference_id}, row_update_params.RowUpdateParams),
),
cast_to=RowStreamResponse,
cast_to=RowUpdateResponse,
)


class RowsResourceWithRawResponse:
def __init__(self, rows: RowsResource) -> None:
self._rows = rows

self.stream = to_raw_response_wrapper(
rows.stream,
self.update = to_raw_response_wrapper(
rows.update,
)


class AsyncRowsResourceWithRawResponse:
def __init__(self, rows: AsyncRowsResource) -> None:
self._rows = rows

self.stream = async_to_raw_response_wrapper(
rows.stream,
self.update = async_to_raw_response_wrapper(
rows.update,
)


class RowsResourceWithStreamingResponse:
def __init__(self, rows: RowsResource) -> None:
self._rows = rows

self.stream = to_streamed_response_wrapper(
rows.stream,
self.update = to_streamed_response_wrapper(
rows.update,
)


class AsyncRowsResourceWithStreamingResponse:
def __init__(self, rows: AsyncRowsResource) -> None:
self._rows = rows

self.stream = async_to_streamed_response_wrapper(
rows.stream,
self.update = async_to_streamed_response_wrapper(
rows.update,
)
4 changes: 2 additions & 2 deletions src/openlayer/types/inference_pipelines/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,9 +2,9 @@

from __future__ import annotations

from .row_stream_params import RowStreamParams as RowStreamParams
from .row_update_params import RowUpdateParams as RowUpdateParams
from .data_stream_params import DataStreamParams as DataStreamParams
from .row_stream_response import RowStreamResponse as RowStreamResponse
from .row_update_response import RowUpdateResponse as RowUpdateResponse
from .data_stream_response import DataStreamResponse as DataStreamResponse
from .test_result_list_params import TestResultListParams as TestResultListParams
from .test_result_list_response import TestResultListResponse as TestResultListResponse
Original file line number Diff line number Diff line change
Expand Up @@ -7,10 +7,10 @@

from ..._utils import PropertyInfo

__all__ = ["RowStreamParams", "Config"]
__all__ = ["RowUpdateParams", "Config"]


class RowStreamParams(TypedDict, total=False):
class RowUpdateParams(TypedDict, total=False):
inference_id: Required[Annotated[str, PropertyInfo(alias="inferenceId")]]
"""Specify the inference id as a query param."""

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@

from ..._models import BaseModel

__all__ = ["RowStreamResponse"]
__all__ = ["RowUpdateResponse"]


class RowStreamResponse(BaseModel):
class RowUpdateResponse(BaseModel):
success: Literal[True]
58 changes: 29 additions & 29 deletions tests/api_resources/inference_pipelines/test_rows.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@

from openlayer import Openlayer, AsyncOpenlayer
from tests.utils import assert_matches_type
from openlayer.types.inference_pipelines import RowStreamResponse
from openlayer.types.inference_pipelines import RowUpdateResponse

base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010")

Expand All @@ -18,17 +18,17 @@ class TestRows:
parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"])

@parametrize
def test_method_stream(self, client: Openlayer) -> None:
row = client.inference_pipelines.rows.stream(
def test_method_update(self, client: Openlayer) -> None:
row = client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
)
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
def test_method_stream_with_all_params(self, client: Openlayer) -> None:
row = client.inference_pipelines.rows.stream(
def test_method_update_with_all_params(self, client: Openlayer) -> None:
row = client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -40,11 +40,11 @@ def test_method_stream_with_all_params(self, client: Openlayer) -> None:
"human_feedback_column_name": "human_feedback",
},
)
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
def test_raw_response_stream(self, client: Openlayer) -> None:
response = client.inference_pipelines.rows.with_raw_response.stream(
def test_raw_response_update(self, client: Openlayer) -> None:
response = client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -53,11 +53,11 @@ def test_raw_response_stream(self, client: Openlayer) -> None:
assert response.is_closed is True
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
row = response.parse()
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
def test_streaming_response_stream(self, client: Openlayer) -> None:
with client.inference_pipelines.rows.with_streaming_response.stream(
def test_streaming_response_update(self, client: Openlayer) -> None:
with client.inference_pipelines.rows.with_streaming_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -66,14 +66,14 @@ def test_streaming_response_stream(self, client: Openlayer) -> None:
assert response.http_request.headers.get("X-Stainless-Lang") == "python"

row = response.parse()
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

assert cast(Any, response.is_closed) is True

@parametrize
def test_path_params_stream(self, client: Openlayer) -> None:
def test_path_params_update(self, client: Openlayer) -> None:
with pytest.raises(ValueError, match=r"Expected a non-empty value for `inference_pipeline_id` but received ''"):
client.inference_pipelines.rows.with_raw_response.stream(
client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="",
inference_id="inferenceId",
row={},
Expand All @@ -84,17 +84,17 @@ class TestAsyncRows:
parametrize = pytest.mark.parametrize("async_client", [False, True], indirect=True, ids=["loose", "strict"])

@parametrize
async def test_method_stream(self, async_client: AsyncOpenlayer) -> None:
row = await async_client.inference_pipelines.rows.stream(
async def test_method_update(self, async_client: AsyncOpenlayer) -> None:
row = await async_client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
)
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
async def test_method_stream_with_all_params(self, async_client: AsyncOpenlayer) -> None:
row = await async_client.inference_pipelines.rows.stream(
async def test_method_update_with_all_params(self, async_client: AsyncOpenlayer) -> None:
row = await async_client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -106,11 +106,11 @@ async def test_method_stream_with_all_params(self, async_client: AsyncOpenlayer)
"human_feedback_column_name": "human_feedback",
},
)
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
async def test_raw_response_stream(self, async_client: AsyncOpenlayer) -> None:
response = await async_client.inference_pipelines.rows.with_raw_response.stream(
async def test_raw_response_update(self, async_client: AsyncOpenlayer) -> None:
response = await async_client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -119,11 +119,11 @@ async def test_raw_response_stream(self, async_client: AsyncOpenlayer) -> None:
assert response.is_closed is True
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
row = await response.parse()
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
async def test_streaming_response_stream(self, async_client: AsyncOpenlayer) -> None:
async with async_client.inference_pipelines.rows.with_streaming_response.stream(
async def test_streaming_response_update(self, async_client: AsyncOpenlayer) -> None:
async with async_client.inference_pipelines.rows.with_streaming_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -132,14 +132,14 @@ async def test_streaming_response_stream(self, async_client: AsyncOpenlayer) ->
assert response.http_request.headers.get("X-Stainless-Lang") == "python"

row = await response.parse()
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

assert cast(Any, response.is_closed) is True

@parametrize
async def test_path_params_stream(self, async_client: AsyncOpenlayer) -> None:
async def test_path_params_update(self, async_client: AsyncOpenlayer) -> None:
with pytest.raises(ValueError, match=r"Expected a non-empty value for `inference_pipeline_id` but received ''"):
await async_client.inference_pipelines.rows.with_raw_response.stream(
await async_client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="",
inference_id="inferenceId",
row={},
Expand Down