diff --git a/api.md b/api.md
index 94c507a4..f9e3456e 100644
--- a/api.md
+++ b/api.md
@@ -69,12 +69,12 @@ Methods:
Types:
```python
-from openlayer.types.inference_pipelines import RowStreamResponse
+from openlayer.types.inference_pipelines import RowUpdateResponse
```
Methods:
-- client.inference_pipelines.rows.stream(inference_pipeline_id, \*\*params) -> RowStreamResponse
+- client.inference_pipelines.rows.update(inference_pipeline_id, \*\*params) -> RowUpdateResponse
## TestResults
diff --git a/src/openlayer/resources/inference_pipelines/rows.py b/src/openlayer/resources/inference_pipelines/rows.py
index f572dc9a..d3407927 100644
--- a/src/openlayer/resources/inference_pipelines/rows.py
+++ b/src/openlayer/resources/inference_pipelines/rows.py
@@ -20,8 +20,8 @@
async_to_streamed_response_wrapper,
)
from ..._base_client import make_request_options
-from ...types.inference_pipelines import row_stream_params
-from ...types.inference_pipelines.row_stream_response import RowStreamResponse
+from ...types.inference_pipelines import row_update_params
+from ...types.inference_pipelines.row_update_response import RowUpdateResponse
__all__ = ["RowsResource", "AsyncRowsResource"]
@@ -35,20 +35,20 @@ def with_raw_response(self) -> RowsResourceWithRawResponse:
def with_streaming_response(self) -> RowsResourceWithStreamingResponse:
return RowsResourceWithStreamingResponse(self)
- def stream(
+ def update(
self,
inference_pipeline_id: str,
*,
inference_id: str,
row: object,
- config: Optional[row_stream_params.Config] | NotGiven = NOT_GIVEN,
+ config: Optional[row_update_params.Config] | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
extra_query: Query | None = None,
extra_body: Body | None = None,
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
- ) -> RowStreamResponse:
+ ) -> RowUpdateResponse:
"""
Update an inference data point in an inference pipeline.
@@ -74,16 +74,16 @@ def stream(
"row": row,
"config": config,
},
- row_stream_params.RowStreamParams,
+ row_update_params.RowUpdateParams,
),
options=make_request_options(
extra_headers=extra_headers,
extra_query=extra_query,
extra_body=extra_body,
timeout=timeout,
- query=maybe_transform({"inference_id": inference_id}, row_stream_params.RowStreamParams),
+ query=maybe_transform({"inference_id": inference_id}, row_update_params.RowUpdateParams),
),
- cast_to=RowStreamResponse,
+ cast_to=RowUpdateResponse,
)
@@ -96,20 +96,20 @@ def with_raw_response(self) -> AsyncRowsResourceWithRawResponse:
def with_streaming_response(self) -> AsyncRowsResourceWithStreamingResponse:
return AsyncRowsResourceWithStreamingResponse(self)
- async def stream(
+ async def update(
self,
inference_pipeline_id: str,
*,
inference_id: str,
row: object,
- config: Optional[row_stream_params.Config] | NotGiven = NOT_GIVEN,
+ config: Optional[row_update_params.Config] | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
extra_query: Query | None = None,
extra_body: Body | None = None,
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
- ) -> RowStreamResponse:
+ ) -> RowUpdateResponse:
"""
Update an inference data point in an inference pipeline.
@@ -135,16 +135,16 @@ async def stream(
"row": row,
"config": config,
},
- row_stream_params.RowStreamParams,
+ row_update_params.RowUpdateParams,
),
options=make_request_options(
extra_headers=extra_headers,
extra_query=extra_query,
extra_body=extra_body,
timeout=timeout,
- query=await async_maybe_transform({"inference_id": inference_id}, row_stream_params.RowStreamParams),
+ query=await async_maybe_transform({"inference_id": inference_id}, row_update_params.RowUpdateParams),
),
- cast_to=RowStreamResponse,
+ cast_to=RowUpdateResponse,
)
@@ -152,8 +152,8 @@ class RowsResourceWithRawResponse:
def __init__(self, rows: RowsResource) -> None:
self._rows = rows
- self.stream = to_raw_response_wrapper(
- rows.stream,
+ self.update = to_raw_response_wrapper(
+ rows.update,
)
@@ -161,8 +161,8 @@ class AsyncRowsResourceWithRawResponse:
def __init__(self, rows: AsyncRowsResource) -> None:
self._rows = rows
- self.stream = async_to_raw_response_wrapper(
- rows.stream,
+ self.update = async_to_raw_response_wrapper(
+ rows.update,
)
@@ -170,8 +170,8 @@ class RowsResourceWithStreamingResponse:
def __init__(self, rows: RowsResource) -> None:
self._rows = rows
- self.stream = to_streamed_response_wrapper(
- rows.stream,
+ self.update = to_streamed_response_wrapper(
+ rows.update,
)
@@ -179,6 +179,6 @@ class AsyncRowsResourceWithStreamingResponse:
def __init__(self, rows: AsyncRowsResource) -> None:
self._rows = rows
- self.stream = async_to_streamed_response_wrapper(
- rows.stream,
+ self.update = async_to_streamed_response_wrapper(
+ rows.update,
)
diff --git a/src/openlayer/types/inference_pipelines/__init__.py b/src/openlayer/types/inference_pipelines/__init__.py
index 93520774..3ccedd4e 100644
--- a/src/openlayer/types/inference_pipelines/__init__.py
+++ b/src/openlayer/types/inference_pipelines/__init__.py
@@ -2,9 +2,9 @@
from __future__ import annotations
-from .row_stream_params import RowStreamParams as RowStreamParams
+from .row_update_params import RowUpdateParams as RowUpdateParams
from .data_stream_params import DataStreamParams as DataStreamParams
-from .row_stream_response import RowStreamResponse as RowStreamResponse
+from .row_update_response import RowUpdateResponse as RowUpdateResponse
from .data_stream_response import DataStreamResponse as DataStreamResponse
from .test_result_list_params import TestResultListParams as TestResultListParams
from .test_result_list_response import TestResultListResponse as TestResultListResponse
diff --git a/src/openlayer/types/inference_pipelines/row_stream_params.py b/src/openlayer/types/inference_pipelines/row_update_params.py
similarity index 94%
rename from src/openlayer/types/inference_pipelines/row_stream_params.py
rename to src/openlayer/types/inference_pipelines/row_update_params.py
index 6e5d1c01..c8af2586 100644
--- a/src/openlayer/types/inference_pipelines/row_stream_params.py
+++ b/src/openlayer/types/inference_pipelines/row_update_params.py
@@ -7,10 +7,10 @@
from ..._utils import PropertyInfo
-__all__ = ["RowStreamParams", "Config"]
+__all__ = ["RowUpdateParams", "Config"]
-class RowStreamParams(TypedDict, total=False):
+class RowUpdateParams(TypedDict, total=False):
inference_id: Required[Annotated[str, PropertyInfo(alias="inferenceId")]]
"""Specify the inference id as a query param."""
diff --git a/src/openlayer/types/inference_pipelines/row_stream_response.py b/src/openlayer/types/inference_pipelines/row_update_response.py
similarity index 73%
rename from src/openlayer/types/inference_pipelines/row_stream_response.py
rename to src/openlayer/types/inference_pipelines/row_update_response.py
index cf65e306..60d9e23d 100644
--- a/src/openlayer/types/inference_pipelines/row_stream_response.py
+++ b/src/openlayer/types/inference_pipelines/row_update_response.py
@@ -4,8 +4,8 @@
from ..._models import BaseModel
-__all__ = ["RowStreamResponse"]
+__all__ = ["RowUpdateResponse"]
-class RowStreamResponse(BaseModel):
+class RowUpdateResponse(BaseModel):
success: Literal[True]
diff --git a/tests/api_resources/inference_pipelines/test_rows.py b/tests/api_resources/inference_pipelines/test_rows.py
index 9e6ace63..a6c95710 100644
--- a/tests/api_resources/inference_pipelines/test_rows.py
+++ b/tests/api_resources/inference_pipelines/test_rows.py
@@ -9,7 +9,7 @@
from openlayer import Openlayer, AsyncOpenlayer
from tests.utils import assert_matches_type
-from openlayer.types.inference_pipelines import RowStreamResponse
+from openlayer.types.inference_pipelines import RowUpdateResponse
base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010")
@@ -18,17 +18,17 @@ class TestRows:
parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"])
@parametrize
- def test_method_stream(self, client: Openlayer) -> None:
- row = client.inference_pipelines.rows.stream(
+ def test_method_update(self, client: Openlayer) -> None:
+ row = client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
)
- assert_matches_type(RowStreamResponse, row, path=["response"])
+ assert_matches_type(RowUpdateResponse, row, path=["response"])
@parametrize
- def test_method_stream_with_all_params(self, client: Openlayer) -> None:
- row = client.inference_pipelines.rows.stream(
+ def test_method_update_with_all_params(self, client: Openlayer) -> None:
+ row = client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
@@ -40,11 +40,11 @@ def test_method_stream_with_all_params(self, client: Openlayer) -> None:
"human_feedback_column_name": "human_feedback",
},
)
- assert_matches_type(RowStreamResponse, row, path=["response"])
+ assert_matches_type(RowUpdateResponse, row, path=["response"])
@parametrize
- def test_raw_response_stream(self, client: Openlayer) -> None:
- response = client.inference_pipelines.rows.with_raw_response.stream(
+ def test_raw_response_update(self, client: Openlayer) -> None:
+ response = client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
@@ -53,11 +53,11 @@ def test_raw_response_stream(self, client: Openlayer) -> None:
assert response.is_closed is True
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
row = response.parse()
- assert_matches_type(RowStreamResponse, row, path=["response"])
+ assert_matches_type(RowUpdateResponse, row, path=["response"])
@parametrize
- def test_streaming_response_stream(self, client: Openlayer) -> None:
- with client.inference_pipelines.rows.with_streaming_response.stream(
+ def test_streaming_response_update(self, client: Openlayer) -> None:
+ with client.inference_pipelines.rows.with_streaming_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
@@ -66,14 +66,14 @@ def test_streaming_response_stream(self, client: Openlayer) -> None:
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
row = response.parse()
- assert_matches_type(RowStreamResponse, row, path=["response"])
+ assert_matches_type(RowUpdateResponse, row, path=["response"])
assert cast(Any, response.is_closed) is True
@parametrize
- def test_path_params_stream(self, client: Openlayer) -> None:
+ def test_path_params_update(self, client: Openlayer) -> None:
with pytest.raises(ValueError, match=r"Expected a non-empty value for `inference_pipeline_id` but received ''"):
- client.inference_pipelines.rows.with_raw_response.stream(
+ client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="",
inference_id="inferenceId",
row={},
@@ -84,17 +84,17 @@ class TestAsyncRows:
parametrize = pytest.mark.parametrize("async_client", [False, True], indirect=True, ids=["loose", "strict"])
@parametrize
- async def test_method_stream(self, async_client: AsyncOpenlayer) -> None:
- row = await async_client.inference_pipelines.rows.stream(
+ async def test_method_update(self, async_client: AsyncOpenlayer) -> None:
+ row = await async_client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
)
- assert_matches_type(RowStreamResponse, row, path=["response"])
+ assert_matches_type(RowUpdateResponse, row, path=["response"])
@parametrize
- async def test_method_stream_with_all_params(self, async_client: AsyncOpenlayer) -> None:
- row = await async_client.inference_pipelines.rows.stream(
+ async def test_method_update_with_all_params(self, async_client: AsyncOpenlayer) -> None:
+ row = await async_client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
@@ -106,11 +106,11 @@ async def test_method_stream_with_all_params(self, async_client: AsyncOpenlayer)
"human_feedback_column_name": "human_feedback",
},
)
- assert_matches_type(RowStreamResponse, row, path=["response"])
+ assert_matches_type(RowUpdateResponse, row, path=["response"])
@parametrize
- async def test_raw_response_stream(self, async_client: AsyncOpenlayer) -> None:
- response = await async_client.inference_pipelines.rows.with_raw_response.stream(
+ async def test_raw_response_update(self, async_client: AsyncOpenlayer) -> None:
+ response = await async_client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
@@ -119,11 +119,11 @@ async def test_raw_response_stream(self, async_client: AsyncOpenlayer) -> None:
assert response.is_closed is True
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
row = await response.parse()
- assert_matches_type(RowStreamResponse, row, path=["response"])
+ assert_matches_type(RowUpdateResponse, row, path=["response"])
@parametrize
- async def test_streaming_response_stream(self, async_client: AsyncOpenlayer) -> None:
- async with async_client.inference_pipelines.rows.with_streaming_response.stream(
+ async def test_streaming_response_update(self, async_client: AsyncOpenlayer) -> None:
+ async with async_client.inference_pipelines.rows.with_streaming_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
@@ -132,14 +132,14 @@ async def test_streaming_response_stream(self, async_client: AsyncOpenlayer) ->
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
row = await response.parse()
- assert_matches_type(RowStreamResponse, row, path=["response"])
+ assert_matches_type(RowUpdateResponse, row, path=["response"])
assert cast(Any, response.is_closed) is True
@parametrize
- async def test_path_params_stream(self, async_client: AsyncOpenlayer) -> None:
+ async def test_path_params_update(self, async_client: AsyncOpenlayer) -> None:
with pytest.raises(ValueError, match=r"Expected a non-empty value for `inference_pipeline_id` but received ''"):
- await async_client.inference_pipelines.rows.with_raw_response.stream(
+ await async_client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="",
inference_id="inferenceId",
row={},