From e01a9628f212e2026277a4cb27020a873c63fbfe Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 17 Jul 2024 21:53:31 +0000 Subject: [PATCH] feat(api): update via SDK Studio (#263) --- api.md | 4 +- .../resources/inference_pipelines/rows.py | 44 +++++++------- .../types/inference_pipelines/__init__.py | 4 +- ..._stream_params.py => row_update_params.py} | 4 +- ...eam_response.py => row_update_response.py} | 4 +- .../inference_pipelines/test_rows.py | 58 +++++++++---------- 6 files changed, 59 insertions(+), 59 deletions(-) rename src/openlayer/types/inference_pipelines/{row_stream_params.py => row_update_params.py} (94%) rename src/openlayer/types/inference_pipelines/{row_stream_response.py => row_update_response.py} (73%) diff --git a/api.md b/api.md index 94c507a4..f9e3456e 100644 --- a/api.md +++ b/api.md @@ -69,12 +69,12 @@ Methods: Types: ```python -from openlayer.types.inference_pipelines import RowStreamResponse +from openlayer.types.inference_pipelines import RowUpdateResponse ``` Methods: -- client.inference_pipelines.rows.stream(inference_pipeline_id, \*\*params) -> RowStreamResponse +- client.inference_pipelines.rows.update(inference_pipeline_id, \*\*params) -> RowUpdateResponse ## TestResults diff --git a/src/openlayer/resources/inference_pipelines/rows.py b/src/openlayer/resources/inference_pipelines/rows.py index f572dc9a..d3407927 100644 --- a/src/openlayer/resources/inference_pipelines/rows.py +++ b/src/openlayer/resources/inference_pipelines/rows.py @@ -20,8 +20,8 @@ async_to_streamed_response_wrapper, ) from ..._base_client import make_request_options -from ...types.inference_pipelines import row_stream_params -from ...types.inference_pipelines.row_stream_response import RowStreamResponse +from ...types.inference_pipelines import row_update_params +from ...types.inference_pipelines.row_update_response import RowUpdateResponse __all__ = ["RowsResource", "AsyncRowsResource"] @@ -35,20 +35,20 @@ def with_raw_response(self) -> RowsResourceWithRawResponse: def with_streaming_response(self) -> RowsResourceWithStreamingResponse: return RowsResourceWithStreamingResponse(self) - def stream( + def update( self, inference_pipeline_id: str, *, inference_id: str, row: object, - config: Optional[row_stream_params.Config] | NotGiven = NOT_GIVEN, + config: Optional[row_update_params.Config] | NotGiven = NOT_GIVEN, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. # The extra values given here take precedence over values defined on the client or passed to this method. extra_headers: Headers | None = None, extra_query: Query | None = None, extra_body: Body | None = None, timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN, - ) -> RowStreamResponse: + ) -> RowUpdateResponse: """ Update an inference data point in an inference pipeline. @@ -74,16 +74,16 @@ def stream( "row": row, "config": config, }, - row_stream_params.RowStreamParams, + row_update_params.RowUpdateParams, ), options=make_request_options( extra_headers=extra_headers, extra_query=extra_query, extra_body=extra_body, timeout=timeout, - query=maybe_transform({"inference_id": inference_id}, row_stream_params.RowStreamParams), + query=maybe_transform({"inference_id": inference_id}, row_update_params.RowUpdateParams), ), - cast_to=RowStreamResponse, + cast_to=RowUpdateResponse, ) @@ -96,20 +96,20 @@ def with_raw_response(self) -> AsyncRowsResourceWithRawResponse: def with_streaming_response(self) -> AsyncRowsResourceWithStreamingResponse: return AsyncRowsResourceWithStreamingResponse(self) - async def stream( + async def update( self, inference_pipeline_id: str, *, inference_id: str, row: object, - config: Optional[row_stream_params.Config] | NotGiven = NOT_GIVEN, + config: Optional[row_update_params.Config] | NotGiven = NOT_GIVEN, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. # The extra values given here take precedence over values defined on the client or passed to this method. extra_headers: Headers | None = None, extra_query: Query | None = None, extra_body: Body | None = None, timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN, - ) -> RowStreamResponse: + ) -> RowUpdateResponse: """ Update an inference data point in an inference pipeline. @@ -135,16 +135,16 @@ async def stream( "row": row, "config": config, }, - row_stream_params.RowStreamParams, + row_update_params.RowUpdateParams, ), options=make_request_options( extra_headers=extra_headers, extra_query=extra_query, extra_body=extra_body, timeout=timeout, - query=await async_maybe_transform({"inference_id": inference_id}, row_stream_params.RowStreamParams), + query=await async_maybe_transform({"inference_id": inference_id}, row_update_params.RowUpdateParams), ), - cast_to=RowStreamResponse, + cast_to=RowUpdateResponse, ) @@ -152,8 +152,8 @@ class RowsResourceWithRawResponse: def __init__(self, rows: RowsResource) -> None: self._rows = rows - self.stream = to_raw_response_wrapper( - rows.stream, + self.update = to_raw_response_wrapper( + rows.update, ) @@ -161,8 +161,8 @@ class AsyncRowsResourceWithRawResponse: def __init__(self, rows: AsyncRowsResource) -> None: self._rows = rows - self.stream = async_to_raw_response_wrapper( - rows.stream, + self.update = async_to_raw_response_wrapper( + rows.update, ) @@ -170,8 +170,8 @@ class RowsResourceWithStreamingResponse: def __init__(self, rows: RowsResource) -> None: self._rows = rows - self.stream = to_streamed_response_wrapper( - rows.stream, + self.update = to_streamed_response_wrapper( + rows.update, ) @@ -179,6 +179,6 @@ class AsyncRowsResourceWithStreamingResponse: def __init__(self, rows: AsyncRowsResource) -> None: self._rows = rows - self.stream = async_to_streamed_response_wrapper( - rows.stream, + self.update = async_to_streamed_response_wrapper( + rows.update, ) diff --git a/src/openlayer/types/inference_pipelines/__init__.py b/src/openlayer/types/inference_pipelines/__init__.py index 93520774..3ccedd4e 100644 --- a/src/openlayer/types/inference_pipelines/__init__.py +++ b/src/openlayer/types/inference_pipelines/__init__.py @@ -2,9 +2,9 @@ from __future__ import annotations -from .row_stream_params import RowStreamParams as RowStreamParams +from .row_update_params import RowUpdateParams as RowUpdateParams from .data_stream_params import DataStreamParams as DataStreamParams -from .row_stream_response import RowStreamResponse as RowStreamResponse +from .row_update_response import RowUpdateResponse as RowUpdateResponse from .data_stream_response import DataStreamResponse as DataStreamResponse from .test_result_list_params import TestResultListParams as TestResultListParams from .test_result_list_response import TestResultListResponse as TestResultListResponse diff --git a/src/openlayer/types/inference_pipelines/row_stream_params.py b/src/openlayer/types/inference_pipelines/row_update_params.py similarity index 94% rename from src/openlayer/types/inference_pipelines/row_stream_params.py rename to src/openlayer/types/inference_pipelines/row_update_params.py index 6e5d1c01..c8af2586 100644 --- a/src/openlayer/types/inference_pipelines/row_stream_params.py +++ b/src/openlayer/types/inference_pipelines/row_update_params.py @@ -7,10 +7,10 @@ from ..._utils import PropertyInfo -__all__ = ["RowStreamParams", "Config"] +__all__ = ["RowUpdateParams", "Config"] -class RowStreamParams(TypedDict, total=False): +class RowUpdateParams(TypedDict, total=False): inference_id: Required[Annotated[str, PropertyInfo(alias="inferenceId")]] """Specify the inference id as a query param.""" diff --git a/src/openlayer/types/inference_pipelines/row_stream_response.py b/src/openlayer/types/inference_pipelines/row_update_response.py similarity index 73% rename from src/openlayer/types/inference_pipelines/row_stream_response.py rename to src/openlayer/types/inference_pipelines/row_update_response.py index cf65e306..60d9e23d 100644 --- a/src/openlayer/types/inference_pipelines/row_stream_response.py +++ b/src/openlayer/types/inference_pipelines/row_update_response.py @@ -4,8 +4,8 @@ from ..._models import BaseModel -__all__ = ["RowStreamResponse"] +__all__ = ["RowUpdateResponse"] -class RowStreamResponse(BaseModel): +class RowUpdateResponse(BaseModel): success: Literal[True] diff --git a/tests/api_resources/inference_pipelines/test_rows.py b/tests/api_resources/inference_pipelines/test_rows.py index 9e6ace63..a6c95710 100644 --- a/tests/api_resources/inference_pipelines/test_rows.py +++ b/tests/api_resources/inference_pipelines/test_rows.py @@ -9,7 +9,7 @@ from openlayer import Openlayer, AsyncOpenlayer from tests.utils import assert_matches_type -from openlayer.types.inference_pipelines import RowStreamResponse +from openlayer.types.inference_pipelines import RowUpdateResponse base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") @@ -18,17 +18,17 @@ class TestRows: parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) @parametrize - def test_method_stream(self, client: Openlayer) -> None: - row = client.inference_pipelines.rows.stream( + def test_method_update(self, client: Openlayer) -> None: + row = client.inference_pipelines.rows.update( inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e", inference_id="inferenceId", row={}, ) - assert_matches_type(RowStreamResponse, row, path=["response"]) + assert_matches_type(RowUpdateResponse, row, path=["response"]) @parametrize - def test_method_stream_with_all_params(self, client: Openlayer) -> None: - row = client.inference_pipelines.rows.stream( + def test_method_update_with_all_params(self, client: Openlayer) -> None: + row = client.inference_pipelines.rows.update( inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e", inference_id="inferenceId", row={}, @@ -40,11 +40,11 @@ def test_method_stream_with_all_params(self, client: Openlayer) -> None: "human_feedback_column_name": "human_feedback", }, ) - assert_matches_type(RowStreamResponse, row, path=["response"]) + assert_matches_type(RowUpdateResponse, row, path=["response"]) @parametrize - def test_raw_response_stream(self, client: Openlayer) -> None: - response = client.inference_pipelines.rows.with_raw_response.stream( + def test_raw_response_update(self, client: Openlayer) -> None: + response = client.inference_pipelines.rows.with_raw_response.update( inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e", inference_id="inferenceId", row={}, @@ -53,11 +53,11 @@ def test_raw_response_stream(self, client: Openlayer) -> None: assert response.is_closed is True assert response.http_request.headers.get("X-Stainless-Lang") == "python" row = response.parse() - assert_matches_type(RowStreamResponse, row, path=["response"]) + assert_matches_type(RowUpdateResponse, row, path=["response"]) @parametrize - def test_streaming_response_stream(self, client: Openlayer) -> None: - with client.inference_pipelines.rows.with_streaming_response.stream( + def test_streaming_response_update(self, client: Openlayer) -> None: + with client.inference_pipelines.rows.with_streaming_response.update( inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e", inference_id="inferenceId", row={}, @@ -66,14 +66,14 @@ def test_streaming_response_stream(self, client: Openlayer) -> None: assert response.http_request.headers.get("X-Stainless-Lang") == "python" row = response.parse() - assert_matches_type(RowStreamResponse, row, path=["response"]) + assert_matches_type(RowUpdateResponse, row, path=["response"]) assert cast(Any, response.is_closed) is True @parametrize - def test_path_params_stream(self, client: Openlayer) -> None: + def test_path_params_update(self, client: Openlayer) -> None: with pytest.raises(ValueError, match=r"Expected a non-empty value for `inference_pipeline_id` but received ''"): - client.inference_pipelines.rows.with_raw_response.stream( + client.inference_pipelines.rows.with_raw_response.update( inference_pipeline_id="", inference_id="inferenceId", row={}, @@ -84,17 +84,17 @@ class TestAsyncRows: parametrize = pytest.mark.parametrize("async_client", [False, True], indirect=True, ids=["loose", "strict"]) @parametrize - async def test_method_stream(self, async_client: AsyncOpenlayer) -> None: - row = await async_client.inference_pipelines.rows.stream( + async def test_method_update(self, async_client: AsyncOpenlayer) -> None: + row = await async_client.inference_pipelines.rows.update( inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e", inference_id="inferenceId", row={}, ) - assert_matches_type(RowStreamResponse, row, path=["response"]) + assert_matches_type(RowUpdateResponse, row, path=["response"]) @parametrize - async def test_method_stream_with_all_params(self, async_client: AsyncOpenlayer) -> None: - row = await async_client.inference_pipelines.rows.stream( + async def test_method_update_with_all_params(self, async_client: AsyncOpenlayer) -> None: + row = await async_client.inference_pipelines.rows.update( inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e", inference_id="inferenceId", row={}, @@ -106,11 +106,11 @@ async def test_method_stream_with_all_params(self, async_client: AsyncOpenlayer) "human_feedback_column_name": "human_feedback", }, ) - assert_matches_type(RowStreamResponse, row, path=["response"]) + assert_matches_type(RowUpdateResponse, row, path=["response"]) @parametrize - async def test_raw_response_stream(self, async_client: AsyncOpenlayer) -> None: - response = await async_client.inference_pipelines.rows.with_raw_response.stream( + async def test_raw_response_update(self, async_client: AsyncOpenlayer) -> None: + response = await async_client.inference_pipelines.rows.with_raw_response.update( inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e", inference_id="inferenceId", row={}, @@ -119,11 +119,11 @@ async def test_raw_response_stream(self, async_client: AsyncOpenlayer) -> None: assert response.is_closed is True assert response.http_request.headers.get("X-Stainless-Lang") == "python" row = await response.parse() - assert_matches_type(RowStreamResponse, row, path=["response"]) + assert_matches_type(RowUpdateResponse, row, path=["response"]) @parametrize - async def test_streaming_response_stream(self, async_client: AsyncOpenlayer) -> None: - async with async_client.inference_pipelines.rows.with_streaming_response.stream( + async def test_streaming_response_update(self, async_client: AsyncOpenlayer) -> None: + async with async_client.inference_pipelines.rows.with_streaming_response.update( inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e", inference_id="inferenceId", row={}, @@ -132,14 +132,14 @@ async def test_streaming_response_stream(self, async_client: AsyncOpenlayer) -> assert response.http_request.headers.get("X-Stainless-Lang") == "python" row = await response.parse() - assert_matches_type(RowStreamResponse, row, path=["response"]) + assert_matches_type(RowUpdateResponse, row, path=["response"]) assert cast(Any, response.is_closed) is True @parametrize - async def test_path_params_stream(self, async_client: AsyncOpenlayer) -> None: + async def test_path_params_update(self, async_client: AsyncOpenlayer) -> None: with pytest.raises(ValueError, match=r"Expected a non-empty value for `inference_pipeline_id` but received ''"): - await async_client.inference_pipelines.rows.with_raw_response.stream( + await async_client.inference_pipelines.rows.with_raw_response.update( inference_pipeline_id="", inference_id="inferenceId", row={},