Skip to content

Commit

Permalink
feat(api): update via SDK Studio (#263)
Browse files Browse the repository at this point in the history
  • Loading branch information
stainless-app[bot] committed Jul 17, 2024
1 parent 65f7548 commit e01a962
Show file tree
Hide file tree
Showing 6 changed files with 59 additions and 59 deletions.
4 changes: 2 additions & 2 deletions api.md
Original file line number Diff line number Diff line change
Expand Up @@ -69,12 +69,12 @@ Methods:
Types:

```python
from openlayer.types.inference_pipelines import RowStreamResponse
from openlayer.types.inference_pipelines import RowUpdateResponse
```

Methods:

- <code title="put /inference-pipelines/{inferencePipelineId}/rows">client.inference_pipelines.rows.<a href="./src/openlayer/resources/inference_pipelines/rows.py">stream</a>(inference_pipeline_id, \*\*<a href="src/openlayer/types/inference_pipelines/row_stream_params.py">params</a>) -> <a href="./src/openlayer/types/inference_pipelines/row_stream_response.py">RowStreamResponse</a></code>
- <code title="put /inference-pipelines/{inferencePipelineId}/rows">client.inference_pipelines.rows.<a href="./src/openlayer/resources/inference_pipelines/rows.py">update</a>(inference_pipeline_id, \*\*<a href="src/openlayer/types/inference_pipelines/row_update_params.py">params</a>) -> <a href="./src/openlayer/types/inference_pipelines/row_update_response.py">RowUpdateResponse</a></code>

## TestResults

Expand Down
44 changes: 22 additions & 22 deletions src/openlayer/resources/inference_pipelines/rows.py
Original file line number Diff line number Diff line change
Expand Up @@ -20,8 +20,8 @@
async_to_streamed_response_wrapper,
)
from ..._base_client import make_request_options
from ...types.inference_pipelines import row_stream_params
from ...types.inference_pipelines.row_stream_response import RowStreamResponse
from ...types.inference_pipelines import row_update_params
from ...types.inference_pipelines.row_update_response import RowUpdateResponse

__all__ = ["RowsResource", "AsyncRowsResource"]

Expand All @@ -35,20 +35,20 @@ def with_raw_response(self) -> RowsResourceWithRawResponse:
def with_streaming_response(self) -> RowsResourceWithStreamingResponse:
return RowsResourceWithStreamingResponse(self)

def stream(
def update(
self,
inference_pipeline_id: str,
*,
inference_id: str,
row: object,
config: Optional[row_stream_params.Config] | NotGiven = NOT_GIVEN,
config: Optional[row_update_params.Config] | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
extra_query: Query | None = None,
extra_body: Body | None = None,
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
) -> RowStreamResponse:
) -> RowUpdateResponse:
"""
Update an inference data point in an inference pipeline.
Expand All @@ -74,16 +74,16 @@ def stream(
"row": row,
"config": config,
},
row_stream_params.RowStreamParams,
row_update_params.RowUpdateParams,
),
options=make_request_options(
extra_headers=extra_headers,
extra_query=extra_query,
extra_body=extra_body,
timeout=timeout,
query=maybe_transform({"inference_id": inference_id}, row_stream_params.RowStreamParams),
query=maybe_transform({"inference_id": inference_id}, row_update_params.RowUpdateParams),
),
cast_to=RowStreamResponse,
cast_to=RowUpdateResponse,
)


Expand All @@ -96,20 +96,20 @@ def with_raw_response(self) -> AsyncRowsResourceWithRawResponse:
def with_streaming_response(self) -> AsyncRowsResourceWithStreamingResponse:
return AsyncRowsResourceWithStreamingResponse(self)

async def stream(
async def update(
self,
inference_pipeline_id: str,
*,
inference_id: str,
row: object,
config: Optional[row_stream_params.Config] | NotGiven = NOT_GIVEN,
config: Optional[row_update_params.Config] | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
# The extra values given here take precedence over values defined on the client or passed to this method.
extra_headers: Headers | None = None,
extra_query: Query | None = None,
extra_body: Body | None = None,
timeout: float | httpx.Timeout | None | NotGiven = NOT_GIVEN,
) -> RowStreamResponse:
) -> RowUpdateResponse:
"""
Update an inference data point in an inference pipeline.
Expand All @@ -135,50 +135,50 @@ async def stream(
"row": row,
"config": config,
},
row_stream_params.RowStreamParams,
row_update_params.RowUpdateParams,
),
options=make_request_options(
extra_headers=extra_headers,
extra_query=extra_query,
extra_body=extra_body,
timeout=timeout,
query=await async_maybe_transform({"inference_id": inference_id}, row_stream_params.RowStreamParams),
query=await async_maybe_transform({"inference_id": inference_id}, row_update_params.RowUpdateParams),
),
cast_to=RowStreamResponse,
cast_to=RowUpdateResponse,
)


class RowsResourceWithRawResponse:
def __init__(self, rows: RowsResource) -> None:
self._rows = rows

self.stream = to_raw_response_wrapper(
rows.stream,
self.update = to_raw_response_wrapper(
rows.update,
)


class AsyncRowsResourceWithRawResponse:
def __init__(self, rows: AsyncRowsResource) -> None:
self._rows = rows

self.stream = async_to_raw_response_wrapper(
rows.stream,
self.update = async_to_raw_response_wrapper(
rows.update,
)


class RowsResourceWithStreamingResponse:
def __init__(self, rows: RowsResource) -> None:
self._rows = rows

self.stream = to_streamed_response_wrapper(
rows.stream,
self.update = to_streamed_response_wrapper(
rows.update,
)


class AsyncRowsResourceWithStreamingResponse:
def __init__(self, rows: AsyncRowsResource) -> None:
self._rows = rows

self.stream = async_to_streamed_response_wrapper(
rows.stream,
self.update = async_to_streamed_response_wrapper(
rows.update,
)
4 changes: 2 additions & 2 deletions src/openlayer/types/inference_pipelines/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,9 +2,9 @@

from __future__ import annotations

from .row_stream_params import RowStreamParams as RowStreamParams
from .row_update_params import RowUpdateParams as RowUpdateParams
from .data_stream_params import DataStreamParams as DataStreamParams
from .row_stream_response import RowStreamResponse as RowStreamResponse
from .row_update_response import RowUpdateResponse as RowUpdateResponse
from .data_stream_response import DataStreamResponse as DataStreamResponse
from .test_result_list_params import TestResultListParams as TestResultListParams
from .test_result_list_response import TestResultListResponse as TestResultListResponse
Original file line number Diff line number Diff line change
Expand Up @@ -7,10 +7,10 @@

from ..._utils import PropertyInfo

__all__ = ["RowStreamParams", "Config"]
__all__ = ["RowUpdateParams", "Config"]


class RowStreamParams(TypedDict, total=False):
class RowUpdateParams(TypedDict, total=False):
inference_id: Required[Annotated[str, PropertyInfo(alias="inferenceId")]]
"""Specify the inference id as a query param."""

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,8 @@

from ..._models import BaseModel

__all__ = ["RowStreamResponse"]
__all__ = ["RowUpdateResponse"]


class RowStreamResponse(BaseModel):
class RowUpdateResponse(BaseModel):
success: Literal[True]
58 changes: 29 additions & 29 deletions tests/api_resources/inference_pipelines/test_rows.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@

from openlayer import Openlayer, AsyncOpenlayer
from tests.utils import assert_matches_type
from openlayer.types.inference_pipelines import RowStreamResponse
from openlayer.types.inference_pipelines import RowUpdateResponse

base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010")

Expand All @@ -18,17 +18,17 @@ class TestRows:
parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"])

@parametrize
def test_method_stream(self, client: Openlayer) -> None:
row = client.inference_pipelines.rows.stream(
def test_method_update(self, client: Openlayer) -> None:
row = client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
)
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
def test_method_stream_with_all_params(self, client: Openlayer) -> None:
row = client.inference_pipelines.rows.stream(
def test_method_update_with_all_params(self, client: Openlayer) -> None:
row = client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -40,11 +40,11 @@ def test_method_stream_with_all_params(self, client: Openlayer) -> None:
"human_feedback_column_name": "human_feedback",
},
)
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
def test_raw_response_stream(self, client: Openlayer) -> None:
response = client.inference_pipelines.rows.with_raw_response.stream(
def test_raw_response_update(self, client: Openlayer) -> None:
response = client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -53,11 +53,11 @@ def test_raw_response_stream(self, client: Openlayer) -> None:
assert response.is_closed is True
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
row = response.parse()
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
def test_streaming_response_stream(self, client: Openlayer) -> None:
with client.inference_pipelines.rows.with_streaming_response.stream(
def test_streaming_response_update(self, client: Openlayer) -> None:
with client.inference_pipelines.rows.with_streaming_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -66,14 +66,14 @@ def test_streaming_response_stream(self, client: Openlayer) -> None:
assert response.http_request.headers.get("X-Stainless-Lang") == "python"

row = response.parse()
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

assert cast(Any, response.is_closed) is True

@parametrize
def test_path_params_stream(self, client: Openlayer) -> None:
def test_path_params_update(self, client: Openlayer) -> None:
with pytest.raises(ValueError, match=r"Expected a non-empty value for `inference_pipeline_id` but received ''"):
client.inference_pipelines.rows.with_raw_response.stream(
client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="",
inference_id="inferenceId",
row={},
Expand All @@ -84,17 +84,17 @@ class TestAsyncRows:
parametrize = pytest.mark.parametrize("async_client", [False, True], indirect=True, ids=["loose", "strict"])

@parametrize
async def test_method_stream(self, async_client: AsyncOpenlayer) -> None:
row = await async_client.inference_pipelines.rows.stream(
async def test_method_update(self, async_client: AsyncOpenlayer) -> None:
row = await async_client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
)
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
async def test_method_stream_with_all_params(self, async_client: AsyncOpenlayer) -> None:
row = await async_client.inference_pipelines.rows.stream(
async def test_method_update_with_all_params(self, async_client: AsyncOpenlayer) -> None:
row = await async_client.inference_pipelines.rows.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -106,11 +106,11 @@ async def test_method_stream_with_all_params(self, async_client: AsyncOpenlayer)
"human_feedback_column_name": "human_feedback",
},
)
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
async def test_raw_response_stream(self, async_client: AsyncOpenlayer) -> None:
response = await async_client.inference_pipelines.rows.with_raw_response.stream(
async def test_raw_response_update(self, async_client: AsyncOpenlayer) -> None:
response = await async_client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -119,11 +119,11 @@ async def test_raw_response_stream(self, async_client: AsyncOpenlayer) -> None:
assert response.is_closed is True
assert response.http_request.headers.get("X-Stainless-Lang") == "python"
row = await response.parse()
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

@parametrize
async def test_streaming_response_stream(self, async_client: AsyncOpenlayer) -> None:
async with async_client.inference_pipelines.rows.with_streaming_response.stream(
async def test_streaming_response_update(self, async_client: AsyncOpenlayer) -> None:
async with async_client.inference_pipelines.rows.with_streaming_response.update(
inference_pipeline_id="182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e",
inference_id="inferenceId",
row={},
Expand All @@ -132,14 +132,14 @@ async def test_streaming_response_stream(self, async_client: AsyncOpenlayer) ->
assert response.http_request.headers.get("X-Stainless-Lang") == "python"

row = await response.parse()
assert_matches_type(RowStreamResponse, row, path=["response"])
assert_matches_type(RowUpdateResponse, row, path=["response"])

assert cast(Any, response.is_closed) is True

@parametrize
async def test_path_params_stream(self, async_client: AsyncOpenlayer) -> None:
async def test_path_params_update(self, async_client: AsyncOpenlayer) -> None:
with pytest.raises(ValueError, match=r"Expected a non-empty value for `inference_pipeline_id` but received ''"):
await async_client.inference_pipelines.rows.with_raw_response.stream(
await async_client.inference_pipelines.rows.with_raw_response.update(
inference_pipeline_id="",
inference_id="inferenceId",
row={},
Expand Down

0 comments on commit e01a962

Please sign in to comment.