diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 4d0e2d9..e2f90fb 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -21,10 +21,10 @@ jobs: runs-on: ${{ github.repository == 'stainless-sdks/warp-api-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} if: (github.event_name == 'push' || github.event.pull_request.head.repo.fork) && (github.event_name != 'push' || github.event.head_commit.message != 'codegen metadata') steps: - - uses: actions/checkout@v6 + - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2 - name: Install uv - uses: astral-sh/setup-uv@v5 + uses: astral-sh/setup-uv@d4b2f3b6ecc6e67c4457f6d3e41ec42d3d0fcb86 # v5.4.2 with: version: '0.10.2' @@ -43,10 +43,10 @@ jobs: id-token: write runs-on: ${{ github.repository == 'stainless-sdks/warp-api-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} steps: - - uses: actions/checkout@v6 + - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2 - name: Install uv - uses: astral-sh/setup-uv@v5 + uses: astral-sh/setup-uv@d4b2f3b6ecc6e67c4457f6d3e41ec42d3d0fcb86 # v5.4.2 with: version: '0.10.2' @@ -61,7 +61,7 @@ jobs: github.repository == 'stainless-sdks/warp-api-python' && !startsWith(github.ref, 'refs/heads/stl/') id: github-oidc - uses: actions/github-script@v8 + uses: actions/github-script@ed597411d8f924073f98dfc5c65a23a2325f34cd # v8.0.0 with: script: core.setOutput('github_token', await core.getIDToken()); @@ -81,10 +81,10 @@ jobs: runs-on: ${{ github.repository == 'stainless-sdks/warp-api-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} if: github.event_name == 'push' || github.event.pull_request.head.repo.fork steps: - - uses: actions/checkout@v6 + - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2 - name: Install uv - uses: astral-sh/setup-uv@v5 + uses: astral-sh/setup-uv@d4b2f3b6ecc6e67c4457f6d3e41ec42d3d0fcb86 # v5.4.2 with: version: '0.10.2' diff --git a/.github/workflows/publish-pypi.yml b/.github/workflows/publish-pypi.yml index 74160e1..bae8377 100644 --- a/.github/workflows/publish-pypi.yml +++ b/.github/workflows/publish-pypi.yml @@ -17,10 +17,10 @@ jobs: id-token: write steps: - - uses: actions/checkout@v6 + - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2 - name: Install uv - uses: astral-sh/setup-uv@v5 + uses: astral-sh/setup-uv@d4b2f3b6ecc6e67c4457f6d3e41ec42d3d0fcb86 # v5.4.2 with: version: '0.9.13' diff --git a/.github/workflows/release-doctor.yml b/.github/workflows/release-doctor.yml index bb0047d..2f3f96e 100644 --- a/.github/workflows/release-doctor.yml +++ b/.github/workflows/release-doctor.yml @@ -12,7 +12,7 @@ jobs: if: github.repository == 'warpdotdev/oz-sdk-python' && (github.event_name == 'push' || github.event_name == 'workflow_dispatch' || startsWith(github.head_ref, 'release-please') || github.head_ref == 'next') steps: - - uses: actions/checkout@v6 + - uses: actions/checkout@de0fac2e4500dabe0009e67214ff5f5447ce83dd # v6.0.2 - name: Check release environment run: | diff --git a/.release-please-manifest.json b/.release-please-manifest.json index a713055..d52d2b9 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "0.12.0" + ".": "0.13.0" } \ No newline at end of file diff --git a/.stats.yml b/.stats.yml index 2cb0b3b..8e0d8bf 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ -configured_endpoints: 22 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/warp-bnavetta/warp-api-1fecc5f5d6ee664d804b81bd1aa6eec4d3f170ffa788d214fead4f7e95ab9d4e.yml -openapi_spec_hash: 82990b03bd5a93e45bfc79db56ae7fc0 -config_hash: f52e7636f248f25c4ea0b086e7326816 +configured_endpoints: 23 +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/warp-bnavetta/warp-api-b719dd35d7850ee303cdebf54fa3dfddb492a6f578344c2060cfae013b61541c.yml +openapi_spec_hash: 4c21e0d940ef5fc42767be5380571c5d +config_hash: 236823a4936c76818117c16aa5c188df diff --git a/CHANGELOG.md b/CHANGELOG.md index 4431ffa..43d4c99 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,30 @@ # Changelog +## 0.13.0 (2026-05-13) + +Full Changelog: [v0.12.0...v0.13.0](https://github.com/warpdotdev/oz-sdk-python/compare/v0.12.0...v0.13.0) + +### Features + +* Add default harness selection for agents ([8cd06f6](https://github.com/warpdotdev/oz-sdk-python/commit/8cd06f6fcb48444e3861ff130d2fb4fa820e3aa7)) +* Add per-agent AWS Bedrock OIDC inference role (backend) ([e4aadec](https://github.com/warpdotdev/oz-sdk-python/commit/e4aadec916b2bea7e05e0e263b307449002e7f8b)) +* **agents:** add prompt property to agent identity data model ([520c835](https://github.com/warpdotdev/oz-sdk-python/commit/520c8350b5b7850242adc082fe83a7ae666e6da6)) +* **api:** api update ([0976362](https://github.com/warpdotdev/oz-sdk-python/commit/09763624c5e172aa28fa43ed18bc0e391c04331d)) +* **api:** api update ([96d640c](https://github.com/warpdotdev/oz-sdk-python/commit/96d640cd68290e1a915fca51a8629cf9dab4d81d)) +* **api:** api update ([2730eea](https://github.com/warpdotdev/oz-sdk-python/commit/2730eea19d91df998de5ef1b5da4989670a5889a)) +* **api:** api update ([af81ef3](https://github.com/warpdotdev/oz-sdk-python/commit/af81ef3a2d48f378209cf69a3074997cb02b1b6c)) +* **api:** api update ([99b2d31](https://github.com/warpdotdev/oz-sdk-python/commit/99b2d31ba1a4c2c3d79fc11c5eb8d611c55613b8)) +* Codex auth: API key support. ([cb84f0c](https://github.com/warpdotdev/oz-sdk-python/commit/cb84f0c6c9625d8c0657c728d5fadae065eb37c9)) +* **internal/types:** support eagerly validating pydantic iterators ([a588f0e](https://github.com/warpdotdev/oz-sdk-python/commit/a588f0e5ba27503659b5abf01d0ebc01f652950d)) +* **memory:** agent identity memory store attachments — API layer ([94b5348](https://github.com/warpdotdev/oz-sdk-python/commit/94b5348152c6b0bfb03b0d3887366c4a65e397fb)) +* **memory:** wire memory stores into run pipeline and add listing endpoint ([6bb74c2](https://github.com/warpdotdev/oz-sdk-python/commit/6bb74c2b695cd268fe8466fc6099f082370ba54e)) +* Retrieve memories in third party harnesses ([7689e12](https://github.com/warpdotdev/oz-sdk-python/commit/7689e121d6f22efad3d81828721f8ed900b9cd28)) + + +### Bug Fixes + +* **client:** add missing f-string prefix in file type error message ([17a8e5b](https://github.com/warpdotdev/oz-sdk-python/commit/17a8e5bf17b882a440a067fee1569caf679f8b55)) + ## 0.12.0 (2026-05-07) Full Changelog: [v0.11.0...v0.12.0](https://github.com/warpdotdev/oz-sdk-python/compare/v0.11.0...v0.12.0) diff --git a/api.md b/api.md index bd33474..e82443c 100644 --- a/api.md +++ b/api.md @@ -94,6 +94,7 @@ Methods: - client.agent.agent.update(uid, \*\*params) -> AgentResponse - client.agent.agent.list() -> ListAgentIdentitiesResponse - client.agent.agent.delete(uid) -> None +- client.agent.agent.get(uid) -> AgentResponse ## Sessions diff --git a/pyproject.toml b/pyproject.toml index 367b68a..533a8e4 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "oz-agent-sdk" -version = "0.12.0" +version = "0.13.0" description = "The official Python library for the oz-api API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/oz_agent_sdk/_files.py b/src/oz_agent_sdk/_files.py index 0fdce17..76da9e0 100644 --- a/src/oz_agent_sdk/_files.py +++ b/src/oz_agent_sdk/_files.py @@ -99,7 +99,7 @@ async def async_to_httpx_files(files: RequestFiles | None) -> HttpxRequestFiles elif is_sequence_t(files): files = [(key, await _async_transform_file(file)) for key, file in files] else: - raise TypeError("Unexpected file type input {type(files)}, expected mapping or sequence") + raise TypeError(f"Unexpected file type input {type(files)}, expected mapping or sequence") return files diff --git a/src/oz_agent_sdk/_models.py b/src/oz_agent_sdk/_models.py index e22dd2a..69f41a6 100644 --- a/src/oz_agent_sdk/_models.py +++ b/src/oz_agent_sdk/_models.py @@ -25,7 +25,9 @@ ClassVar, Protocol, Required, + Annotated, ParamSpec, + TypeAlias, TypedDict, TypeGuard, final, @@ -79,7 +81,15 @@ from ._constants import RAW_RESPONSE_HEADER if TYPE_CHECKING: + from pydantic import GetCoreSchemaHandler, ValidatorFunctionWrapHandler + from pydantic_core import CoreSchema, core_schema from pydantic_core.core_schema import ModelField, ModelSchema, LiteralSchema, ModelFieldsSchema +else: + try: + from pydantic_core import CoreSchema, core_schema + except ImportError: + CoreSchema = None + core_schema = None __all__ = ["BaseModel", "GenericModel"] @@ -396,6 +406,76 @@ def model_dump_json( ) +class _EagerIterable(list[_T], Generic[_T]): + """ + Accepts any Iterable[T] input (including generators), consumes it + eagerly, and validates all items upfront. + + Validation preserves the original container type where possible + (e.g. a set[T] stays a set[T]). Serialization (model_dump / JSON) + always emits a list — round-tripping through model_dump() will not + restore the original container type. + """ + + @classmethod + def __get_pydantic_core_schema__( + cls, + source_type: Any, + handler: GetCoreSchemaHandler, + ) -> CoreSchema: + (item_type,) = get_args(source_type) or (Any,) + item_schema: CoreSchema = handler.generate_schema(item_type) + list_of_items_schema: CoreSchema = core_schema.list_schema(item_schema) + + return core_schema.no_info_wrap_validator_function( + cls._validate, + list_of_items_schema, + serialization=core_schema.plain_serializer_function_ser_schema( + cls._serialize, + info_arg=False, + ), + ) + + @staticmethod + def _validate(v: Iterable[_T], handler: "ValidatorFunctionWrapHandler") -> Any: + original_type: type[Any] = type(v) + + # Normalize to list so list_schema can validate each item + if isinstance(v, list): + items: list[_T] = v + else: + try: + items = list(v) + except TypeError as e: + raise TypeError("Value is not iterable") from e + + # Validate items against the inner schema + validated: list[_T] = handler(items) + + # Reconstruct original container type + if original_type is list: + return validated + # str(list) produces the list's repr, not a string built from items, + # so skip reconstruction for str and its subclasses. + if issubclass(original_type, str): + return validated + try: + return original_type(validated) + except (TypeError, ValueError): + # If the type cannot be reconstructed, just return the validated list + return validated + + @staticmethod + def _serialize(v: Iterable[_T]) -> list[_T]: + """Always serialize as a list so Pydantic's JSON encoder is happy.""" + if isinstance(v, list): + return v + return list(v) + + +EagerIterable: TypeAlias = Annotated[Iterable[_T], _EagerIterable] + + def _construct_field(value: object, field: FieldInfo, key: str) -> object: if value is None: return field_get_default(field) diff --git a/src/oz_agent_sdk/_version.py b/src/oz_agent_sdk/_version.py index b96fedf..1762ab1 100644 --- a/src/oz_agent_sdk/_version.py +++ b/src/oz_agent_sdk/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "oz_agent_sdk" -__version__ = "0.12.0" # x-release-please-version +__version__ = "0.13.0" # x-release-please-version diff --git a/src/oz_agent_sdk/resources/agent/agent_.py b/src/oz_agent_sdk/resources/agent/agent_.py index ab7835e..447bc5d 100644 --- a/src/oz_agent_sdk/resources/agent/agent_.py +++ b/src/oz_agent_sdk/resources/agent/agent_.py @@ -50,7 +50,11 @@ def create( self, *, name: str, + base_model: Optional[str] | Omit = omit, description: Optional[str] | Omit = omit, + inference_providers: agent_create_params.InferenceProviders | Omit = omit, + memory_stores: Iterable[agent_create_params.MemoryStore] | Omit = omit, + prompt: Optional[str] | Omit = omit, secrets: Iterable[agent_create_params.Secret] | Omit = omit, skills: SequenceNotStr[str] | Omit = omit, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. @@ -68,8 +72,18 @@ def create( Args: name: A name for the agent + base_model: Optional base model for runs executed by this agent. + description: Optional description of the agent + inference_providers: Inference provider settings used for LLM calls. + + memory_stores: Optional list of memory stores to attach to the agent. Each store must be + team-owned by the same team as the agent. Duplicate UIDs within a single request + are rejected. + + prompt: Optional base prompt for this agent + secrets: Optional list of secrets associated with the agent. Duplicate names within a single request are rejected. Each entry is unioned into the run-time secret scope when the agent executes. @@ -94,7 +108,11 @@ def create( body=maybe_transform( { "name": name, + "base_model": base_model, "description": description, + "inference_providers": inference_providers, + "memory_stores": memory_stores, + "prompt": prompt, "secrets": secrets, "skills": skills, }, @@ -110,8 +128,12 @@ def update( self, uid: str, *, + base_model: Optional[str] | Omit = omit, description: Optional[str] | Omit = omit, + inference_providers: Optional[agent_update_params.InferenceProviders] | Omit = omit, + memory_stores: Optional[Iterable[agent_update_params.MemoryStore]] | Omit = omit, name: str | Omit = omit, + prompt: Optional[str] | Omit = omit, secrets: Optional[Iterable[agent_update_params.Secret]] | Omit = omit, skills: Optional[SequenceNotStr[str]] | Omit = omit, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. @@ -124,13 +146,24 @@ def update( """Update an existing agent. Args: - description: Replacement description. + base_model: Replacement base model. + + Omit or pass `null` to leave unchanged, or pass an empty + string to clear. - Omit or pass `null` to leave unchanged, or use an empty + description: Replacement description. Omit or pass `null` to leave unchanged, or use an empty value to clear. + inference_providers: Inference provider settings used for LLM calls. + + memory_stores: Replacement list of memory stores. Omit to leave unchanged, pass an empty array + to clear, or pass a non-empty array to replace. + name: The new name for the agent + prompt: Replacement prompt. Omit or pass `null` to leave unchanged, or use an empty + value to clear. + secrets: Replacement list of secrets. Omit to leave unchanged, pass an empty array to clear, or pass a non-empty array to replace. Duplicate names are rejected. @@ -151,8 +184,12 @@ def update( path_template("/agent/identities/{uid}", uid=uid), body=maybe_transform( { + "base_model": base_model, "description": description, + "inference_providers": inference_providers, + "memory_stores": memory_stores, "name": name, + "prompt": prompt, "secrets": secrets, "skills": skills, }, @@ -222,6 +259,42 @@ def delete( cast_to=NoneType, ) + def get( + self, + uid: str, + *, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> AgentResponse: + """Retrieve a single agent by its unique identifier. + + The response includes an + `available` flag indicating whether the agent is within the team's plan limit + and may be used for runs. + + Args: + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + if not uid: + raise ValueError(f"Expected a non-empty value for `uid` but received {uid!r}") + return self._get( + path_template("/agent/identities/{uid}", uid=uid), + options=make_request_options( + extra_headers=extra_headers, extra_query=extra_query, extra_body=extra_body, timeout=timeout + ), + cast_to=AgentResponse, + ) + class AsyncAgentResource(AsyncAPIResource): """Operations for running and managing cloud agents""" @@ -249,7 +322,11 @@ async def create( self, *, name: str, + base_model: Optional[str] | Omit = omit, description: Optional[str] | Omit = omit, + inference_providers: agent_create_params.InferenceProviders | Omit = omit, + memory_stores: Iterable[agent_create_params.MemoryStore] | Omit = omit, + prompt: Optional[str] | Omit = omit, secrets: Iterable[agent_create_params.Secret] | Omit = omit, skills: SequenceNotStr[str] | Omit = omit, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. @@ -267,8 +344,18 @@ async def create( Args: name: A name for the agent + base_model: Optional base model for runs executed by this agent. + description: Optional description of the agent + inference_providers: Inference provider settings used for LLM calls. + + memory_stores: Optional list of memory stores to attach to the agent. Each store must be + team-owned by the same team as the agent. Duplicate UIDs within a single request + are rejected. + + prompt: Optional base prompt for this agent + secrets: Optional list of secrets associated with the agent. Duplicate names within a single request are rejected. Each entry is unioned into the run-time secret scope when the agent executes. @@ -293,7 +380,11 @@ async def create( body=await async_maybe_transform( { "name": name, + "base_model": base_model, "description": description, + "inference_providers": inference_providers, + "memory_stores": memory_stores, + "prompt": prompt, "secrets": secrets, "skills": skills, }, @@ -309,8 +400,12 @@ async def update( self, uid: str, *, + base_model: Optional[str] | Omit = omit, description: Optional[str] | Omit = omit, + inference_providers: Optional[agent_update_params.InferenceProviders] | Omit = omit, + memory_stores: Optional[Iterable[agent_update_params.MemoryStore]] | Omit = omit, name: str | Omit = omit, + prompt: Optional[str] | Omit = omit, secrets: Optional[Iterable[agent_update_params.Secret]] | Omit = omit, skills: Optional[SequenceNotStr[str]] | Omit = omit, # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. @@ -323,13 +418,24 @@ async def update( """Update an existing agent. Args: - description: Replacement description. + base_model: Replacement base model. + + Omit or pass `null` to leave unchanged, or pass an empty + string to clear. - Omit or pass `null` to leave unchanged, or use an empty + description: Replacement description. Omit or pass `null` to leave unchanged, or use an empty value to clear. + inference_providers: Inference provider settings used for LLM calls. + + memory_stores: Replacement list of memory stores. Omit to leave unchanged, pass an empty array + to clear, or pass a non-empty array to replace. + name: The new name for the agent + prompt: Replacement prompt. Omit or pass `null` to leave unchanged, or use an empty + value to clear. + secrets: Replacement list of secrets. Omit to leave unchanged, pass an empty array to clear, or pass a non-empty array to replace. Duplicate names are rejected. @@ -350,8 +456,12 @@ async def update( path_template("/agent/identities/{uid}", uid=uid), body=await async_maybe_transform( { + "base_model": base_model, "description": description, + "inference_providers": inference_providers, + "memory_stores": memory_stores, "name": name, + "prompt": prompt, "secrets": secrets, "skills": skills, }, @@ -421,6 +531,42 @@ async def delete( cast_to=NoneType, ) + async def get( + self, + uid: str, + *, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> AgentResponse: + """Retrieve a single agent by its unique identifier. + + The response includes an + `available` flag indicating whether the agent is within the team's plan limit + and may be used for runs. + + Args: + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + if not uid: + raise ValueError(f"Expected a non-empty value for `uid` but received {uid!r}") + return await self._get( + path_template("/agent/identities/{uid}", uid=uid), + options=make_request_options( + extra_headers=extra_headers, extra_query=extra_query, extra_body=extra_body, timeout=timeout + ), + cast_to=AgentResponse, + ) + class AgentResourceWithRawResponse: def __init__(self, agent: AgentResource) -> None: @@ -438,6 +584,9 @@ def __init__(self, agent: AgentResource) -> None: self.delete = to_raw_response_wrapper( agent.delete, ) + self.get = to_raw_response_wrapper( + agent.get, + ) class AsyncAgentResourceWithRawResponse: @@ -456,6 +605,9 @@ def __init__(self, agent: AsyncAgentResource) -> None: self.delete = async_to_raw_response_wrapper( agent.delete, ) + self.get = async_to_raw_response_wrapper( + agent.get, + ) class AgentResourceWithStreamingResponse: @@ -474,6 +626,9 @@ def __init__(self, agent: AgentResource) -> None: self.delete = to_streamed_response_wrapper( agent.delete, ) + self.get = to_streamed_response_wrapper( + agent.get, + ) class AsyncAgentResourceWithStreamingResponse: @@ -492,3 +647,6 @@ def __init__(self, agent: AsyncAgentResource) -> None: self.delete = async_to_streamed_response_wrapper( agent.delete, ) + self.get = async_to_streamed_response_wrapper( + agent.get, + ) diff --git a/src/oz_agent_sdk/types/agent/agent_create_params.py b/src/oz_agent_sdk/types/agent/agent_create_params.py index 5f31e19..db351bc 100644 --- a/src/oz_agent_sdk/types/agent/agent_create_params.py +++ b/src/oz_agent_sdk/types/agent/agent_create_params.py @@ -3,20 +3,36 @@ from __future__ import annotations from typing import Iterable, Optional -from typing_extensions import Required, TypedDict +from typing_extensions import Literal, Required, TypedDict from ..._types import SequenceNotStr -__all__ = ["AgentCreateParams", "Secret"] +__all__ = ["AgentCreateParams", "InferenceProviders", "InferenceProvidersAws", "MemoryStore", "Secret"] class AgentCreateParams(TypedDict, total=False): name: Required[str] """A name for the agent""" + base_model: Optional[str] + """Optional base model for runs executed by this agent.""" + description: Optional[str] """Optional description of the agent""" + inference_providers: InferenceProviders + """Inference provider settings used for LLM calls.""" + + memory_stores: Iterable[MemoryStore] + """ + Optional list of memory stores to attach to the agent. Each store must be + team-owned by the same team as the agent. Duplicate UIDs within a single request + are rejected. + """ + + prompt: Optional[str] + """Optional base prompt for this agent""" + secrets: Iterable[Secret] """ Optional list of secrets associated with the agent. Duplicate names within a @@ -34,6 +50,39 @@ class AgentCreateParams(TypedDict, total=False): """ +class InferenceProvidersAws(TypedDict, total=False): + """ + Configures AWS Bedrock as the LLM inference provider for this + agent or run. + """ + + disabled: bool + """If true, opt out of Bedrock at this layer.""" + + role_arn: str + """IAM role ARN to assume when calling Bedrock.""" + + +class InferenceProviders(TypedDict, total=False): + """Inference provider settings used for LLM calls.""" + + aws: InferenceProvidersAws + """Configures AWS Bedrock as the LLM inference provider for this agent or run.""" + + +class MemoryStore(TypedDict, total=False): + """Reference to a memory store to attach to an agent.""" + + access: Required[Literal["read_write", "read_only"]] + """Access level for the store.""" + + instructions: Required[str] + """Instructions for how the agent should use this memory store. Must not be empty.""" + + uid: Required[str] + """UID of the memory store.""" + + class Secret(TypedDict, total=False): """Reference to a managed secret by name.""" diff --git a/src/oz_agent_sdk/types/agent/agent_response.py b/src/oz_agent_sdk/types/agent/agent_response.py index 8b46c85..fac7f4a 100644 --- a/src/oz_agent_sdk/types/agent/agent_response.py +++ b/src/oz_agent_sdk/types/agent/agent_response.py @@ -2,10 +2,24 @@ from typing import List, Optional from datetime import datetime +from typing_extensions import Literal from ..._models import BaseModel -__all__ = ["AgentResponse", "Secret"] +__all__ = ["AgentResponse", "MemoryStore", "Secret", "InferenceProviders", "InferenceProvidersAws"] + + +class MemoryStore(BaseModel): + """Reference to a memory store to attach to an agent.""" + + access: Literal["read_write", "read_only"] + """Access level for the store.""" + + instructions: str + """Instructions for how the agent should use this memory store. Must not be empty.""" + + uid: str + """UID of the memory store.""" class Secret(BaseModel): @@ -15,6 +29,26 @@ class Secret(BaseModel): """Name of the managed secret.""" +class InferenceProvidersAws(BaseModel): + """ + Configures AWS Bedrock as the LLM inference provider for this + agent or run. + """ + + disabled: Optional[bool] = None + """If true, opt out of Bedrock at this layer.""" + + role_arn: Optional[str] = None + """IAM role ARN to assume when calling Bedrock.""" + + +class InferenceProviders(BaseModel): + """Inference provider settings used for LLM calls.""" + + aws: Optional[InferenceProvidersAws] = None + """Configures AWS Bedrock as the LLM inference provider for this agent or run.""" + + class AgentResponse(BaseModel): available: bool """Whether this agent is within the team's plan limit and can be used for runs""" @@ -22,6 +56,12 @@ class AgentResponse(BaseModel): created_at: datetime """When the agent was created (RFC3339)""" + memory_stores: List[MemoryStore] + """ + Memory stores attached to this agent. Always present; empty when no stores are + attached. + """ + name: str """Name of the agent""" @@ -37,5 +77,21 @@ class AgentResponse(BaseModel): uid: str """Unique identifier for the agent""" + base_model: Optional[str] = None + """Base model for runs executed by this agent. + + The precedence order for model resolution is: + + 1. The model specified on the run itself + 2. The agent's base model + 3. The team's default model + """ + description: Optional[str] = None """Optional description of the agent""" + + inference_providers: Optional[InferenceProviders] = None + """Inference provider settings used for LLM calls.""" + + prompt: Optional[str] = None + """Optional base prompt for this agent""" diff --git a/src/oz_agent_sdk/types/agent/agent_update_params.py b/src/oz_agent_sdk/types/agent/agent_update_params.py index d3cc68a..29278da 100644 --- a/src/oz_agent_sdk/types/agent/agent_update_params.py +++ b/src/oz_agent_sdk/types/agent/agent_update_params.py @@ -3,23 +3,45 @@ from __future__ import annotations from typing import Iterable, Optional -from typing_extensions import Required, TypedDict +from typing_extensions import Literal, Required, TypedDict from ..._types import SequenceNotStr -__all__ = ["AgentUpdateParams", "Secret"] +__all__ = ["AgentUpdateParams", "InferenceProviders", "InferenceProvidersAws", "MemoryStore", "Secret"] class AgentUpdateParams(TypedDict, total=False): + base_model: Optional[str] + """Replacement base model. + + Omit or pass `null` to leave unchanged, or pass an empty string to clear. + """ + description: Optional[str] """Replacement description. Omit or pass `null` to leave unchanged, or use an empty value to clear. """ + inference_providers: Optional[InferenceProviders] + """Inference provider settings used for LLM calls.""" + + memory_stores: Optional[Iterable[MemoryStore]] + """Replacement list of memory stores. + + Omit to leave unchanged, pass an empty array to clear, or pass a non-empty array + to replace. + """ + name: str """The new name for the agent""" + prompt: Optional[str] + """Replacement prompt. + + Omit or pass `null` to leave unchanged, or use an empty value to clear. + """ + secrets: Optional[Iterable[Secret]] """Replacement list of secrets. @@ -35,6 +57,39 @@ class AgentUpdateParams(TypedDict, total=False): """ +class InferenceProvidersAws(TypedDict, total=False): + """ + Configures AWS Bedrock as the LLM inference provider for this + agent or run. + """ + + disabled: bool + """If true, opt out of Bedrock at this layer.""" + + role_arn: str + """IAM role ARN to assume when calling Bedrock.""" + + +class InferenceProviders(TypedDict, total=False): + """Inference provider settings used for LLM calls.""" + + aws: InferenceProvidersAws + """Configures AWS Bedrock as the LLM inference provider for this agent or run.""" + + +class MemoryStore(TypedDict, total=False): + """Reference to a memory store to attach to an agent.""" + + access: Required[Literal["read_write", "read_only"]] + """Access level for the store.""" + + instructions: Required[str] + """Instructions for how the agent should use this memory store. Must not be empty.""" + + uid: Required[str] + """UID of the memory store.""" + + class Secret(TypedDict, total=False): """Reference to a managed secret by name.""" diff --git a/src/oz_agent_sdk/types/agent/run_item.py b/src/oz_agent_sdk/types/agent/run_item.py index 5b802d8..4293ce2 100644 --- a/src/oz_agent_sdk/types/agent/run_item.py +++ b/src/oz_agent_sdk/types/agent/run_item.py @@ -44,6 +44,9 @@ class RequestUsage(BaseModel): inference_cost: Optional[float] = None """Cost of LLM inference for the run""" + platform_cost: Optional[float] = None + """Cost of platform usage for the run""" + class Schedule(BaseModel): """ diff --git a/src/oz_agent_sdk/types/ambient_agent_config.py b/src/oz_agent_sdk/types/ambient_agent_config.py index 148931d..2d5b76a 100644 --- a/src/oz_agent_sdk/types/ambient_agent_config.py +++ b/src/oz_agent_sdk/types/ambient_agent_config.py @@ -1,6 +1,6 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. -from typing import Dict, Optional +from typing import Dict, List, Optional from typing_extensions import Literal from pydantic import Field as FieldInfo @@ -8,7 +8,15 @@ from .._models import BaseModel from .mcp_server_config import McpServerConfig -__all__ = ["AmbientAgentConfig", "Harness", "HarnessAuthSecrets", "SessionSharing"] +__all__ = [ + "AmbientAgentConfig", + "Harness", + "HarnessAuthSecrets", + "InferenceProviders", + "InferenceProvidersAws", + "MemoryStore", + "SessionSharing", +] class Harness(BaseModel): @@ -41,6 +49,39 @@ class HarnessAuthSecrets(BaseModel): """ +class InferenceProvidersAws(BaseModel): + """ + Configures AWS Bedrock as the LLM inference provider for this + agent or run. + """ + + disabled: Optional[bool] = None + """If true, opt out of Bedrock at this layer.""" + + role_arn: Optional[str] = None + """IAM role ARN to assume when calling Bedrock.""" + + +class InferenceProviders(BaseModel): + """Inference provider settings used for LLM calls.""" + + aws: Optional[InferenceProvidersAws] = None + """Configures AWS Bedrock as the LLM inference provider for this agent or run.""" + + +class MemoryStore(BaseModel): + """Reference to a memory store to attach to an agent.""" + + access: Literal["read_write", "read_only"] + """Access level for the store.""" + + instructions: str + """Instructions for how the agent should use this memory store. Must not be empty.""" + + uid: str + """UID of the memory store.""" + + class SessionSharing(BaseModel): """ Configures sharing behavior for the run's shared session. @@ -97,9 +138,15 @@ class AmbientAgentConfig(BaseModel): floor(max_instance_runtime_seconds / 60) for your billing tier). """ + inference_providers: Optional[InferenceProviders] = None + """Inference provider settings used for LLM calls.""" + mcp_servers: Optional[Dict[str, McpServerConfig]] = None """Map of MCP server configurations by name""" + memory_stores: Optional[List[MemoryStore]] = None + """Memory stores to attach to this run.""" + api_model_id: Optional[str] = FieldInfo(alias="model_id", default=None) """LLM model to use (uses team default if not specified)""" diff --git a/src/oz_agent_sdk/types/ambient_agent_config_param.py b/src/oz_agent_sdk/types/ambient_agent_config_param.py index f4c6ee4..4e662df 100644 --- a/src/oz_agent_sdk/types/ambient_agent_config_param.py +++ b/src/oz_agent_sdk/types/ambient_agent_config_param.py @@ -2,12 +2,20 @@ from __future__ import annotations -from typing import Dict -from typing_extensions import Literal, TypedDict +from typing import Dict, Iterable +from typing_extensions import Literal, Required, TypedDict from .mcp_server_config_param import McpServerConfigParam -__all__ = ["AmbientAgentConfigParam", "Harness", "HarnessAuthSecrets", "SessionSharing"] +__all__ = [ + "AmbientAgentConfigParam", + "Harness", + "HarnessAuthSecrets", + "InferenceProviders", + "InferenceProvidersAws", + "MemoryStore", + "SessionSharing", +] class Harness(TypedDict, total=False): @@ -40,6 +48,39 @@ class HarnessAuthSecrets(TypedDict, total=False): """ +class InferenceProvidersAws(TypedDict, total=False): + """ + Configures AWS Bedrock as the LLM inference provider for this + agent or run. + """ + + disabled: bool + """If true, opt out of Bedrock at this layer.""" + + role_arn: str + """IAM role ARN to assume when calling Bedrock.""" + + +class InferenceProviders(TypedDict, total=False): + """Inference provider settings used for LLM calls.""" + + aws: InferenceProvidersAws + """Configures AWS Bedrock as the LLM inference provider for this agent or run.""" + + +class MemoryStore(TypedDict, total=False): + """Reference to a memory store to attach to an agent.""" + + access: Required[Literal["read_write", "read_only"]] + """Access level for the store.""" + + instructions: Required[str] + """Instructions for how the agent should use this memory store. Must not be empty.""" + + uid: Required[str] + """UID of the memory store.""" + + class SessionSharing(TypedDict, total=False): """ Configures sharing behavior for the run's shared session. @@ -96,9 +137,15 @@ class AmbientAgentConfigParam(TypedDict, total=False): floor(max_instance_runtime_seconds / 60) for your billing tier). """ + inference_providers: InferenceProviders + """Inference provider settings used for LLM calls.""" + mcp_servers: Dict[str, McpServerConfigParam] """Map of MCP server configurations by name""" + memory_stores: Iterable[MemoryStore] + """Memory stores to attach to this run.""" + model_id: str """LLM model to use (uses team default if not specified)""" diff --git a/tests/api_resources/agent/test_agent_.py b/tests/api_resources/agent/test_agent_.py index a23b05e..0b91ac0 100644 --- a/tests/api_resources/agent/test_agent_.py +++ b/tests/api_resources/agent/test_agent_.py @@ -33,7 +33,22 @@ def test_method_create(self, client: OzAPI) -> None: def test_method_create_with_all_params(self, client: OzAPI) -> None: agent = client.agent.agent.create( name="name", + base_model="base_model", description="description", + inference_providers={ + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, + memory_stores=[ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], + prompt="prompt", secrets=[{"name": "name"}], skills=["string"], ) @@ -78,8 +93,23 @@ def test_method_update(self, client: OzAPI) -> None: def test_method_update_with_all_params(self, client: OzAPI) -> None: agent = client.agent.agent.update( uid="uid", + base_model="base_model", description="description", + inference_providers={ + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, + memory_stores=[ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], name="name", + prompt="prompt", secrets=[{"name": "name"}], skills=["string"], ) @@ -189,6 +219,48 @@ def test_path_params_delete(self, client: OzAPI) -> None: "", ) + @pytest.mark.skip(reason="Mock server tests are disabled") + @parametrize + def test_method_get(self, client: OzAPI) -> None: + agent = client.agent.agent.get( + "uid", + ) + assert_matches_type(AgentResponse, agent, path=["response"]) + + @pytest.mark.skip(reason="Mock server tests are disabled") + @parametrize + def test_raw_response_get(self, client: OzAPI) -> None: + response = client.agent.agent.with_raw_response.get( + "uid", + ) + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + agent = response.parse() + assert_matches_type(AgentResponse, agent, path=["response"]) + + @pytest.mark.skip(reason="Mock server tests are disabled") + @parametrize + def test_streaming_response_get(self, client: OzAPI) -> None: + with client.agent.agent.with_streaming_response.get( + "uid", + ) as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + agent = response.parse() + assert_matches_type(AgentResponse, agent, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Mock server tests are disabled") + @parametrize + def test_path_params_get(self, client: OzAPI) -> None: + with pytest.raises(ValueError, match=r"Expected a non-empty value for `uid` but received ''"): + client.agent.agent.with_raw_response.get( + "", + ) + class TestAsyncAgent: parametrize = pytest.mark.parametrize( @@ -208,7 +280,22 @@ async def test_method_create(self, async_client: AsyncOzAPI) -> None: async def test_method_create_with_all_params(self, async_client: AsyncOzAPI) -> None: agent = await async_client.agent.agent.create( name="name", + base_model="base_model", description="description", + inference_providers={ + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, + memory_stores=[ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], + prompt="prompt", secrets=[{"name": "name"}], skills=["string"], ) @@ -253,8 +340,23 @@ async def test_method_update(self, async_client: AsyncOzAPI) -> None: async def test_method_update_with_all_params(self, async_client: AsyncOzAPI) -> None: agent = await async_client.agent.agent.update( uid="uid", + base_model="base_model", description="description", + inference_providers={ + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, + memory_stores=[ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], name="name", + prompt="prompt", secrets=[{"name": "name"}], skills=["string"], ) @@ -363,3 +465,45 @@ async def test_path_params_delete(self, async_client: AsyncOzAPI) -> None: await async_client.agent.agent.with_raw_response.delete( "", ) + + @pytest.mark.skip(reason="Mock server tests are disabled") + @parametrize + async def test_method_get(self, async_client: AsyncOzAPI) -> None: + agent = await async_client.agent.agent.get( + "uid", + ) + assert_matches_type(AgentResponse, agent, path=["response"]) + + @pytest.mark.skip(reason="Mock server tests are disabled") + @parametrize + async def test_raw_response_get(self, async_client: AsyncOzAPI) -> None: + response = await async_client.agent.agent.with_raw_response.get( + "uid", + ) + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + agent = await response.parse() + assert_matches_type(AgentResponse, agent, path=["response"]) + + @pytest.mark.skip(reason="Mock server tests are disabled") + @parametrize + async def test_streaming_response_get(self, async_client: AsyncOzAPI) -> None: + async with async_client.agent.agent.with_streaming_response.get( + "uid", + ) as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + agent = await response.parse() + assert_matches_type(AgentResponse, agent, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Mock server tests are disabled") + @parametrize + async def test_path_params_get(self, async_client: AsyncOzAPI) -> None: + with pytest.raises(ValueError, match=r"Expected a non-empty value for `uid` but received ''"): + await async_client.agent.agent.with_raw_response.get( + "", + ) diff --git a/tests/api_resources/agent/test_schedules.py b/tests/api_resources/agent/test_schedules.py index 0bf7d09..38f62c2 100644 --- a/tests/api_resources/agent/test_schedules.py +++ b/tests/api_resources/agent/test_schedules.py @@ -43,6 +43,12 @@ def test_method_create_with_all_params(self, client: OzAPI) -> None: "harness": {"type": "oz"}, "harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"}, "idle_timeout_minutes": 1, + "inference_providers": { + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, "mcp_servers": { "foo": { "args": ["string"], @@ -53,6 +59,13 @@ def test_method_create_with_all_params(self, client: OzAPI) -> None: "warp_id": "warp_id", } }, + "memory_stores": [ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], "model_id": "model_id", "name": "name", "session_sharing": {"public_access": "VIEWER"}, @@ -163,6 +176,12 @@ def test_method_update_with_all_params(self, client: OzAPI) -> None: "harness": {"type": "oz"}, "harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"}, "idle_timeout_minutes": 1, + "inference_providers": { + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, "mcp_servers": { "foo": { "args": ["string"], @@ -173,6 +192,13 @@ def test_method_update_with_all_params(self, client: OzAPI) -> None: "warp_id": "warp_id", } }, + "memory_stores": [ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], "model_id": "model_id", "name": "name", "session_sharing": {"public_access": "VIEWER"}, @@ -410,6 +436,12 @@ async def test_method_create_with_all_params(self, async_client: AsyncOzAPI) -> "harness": {"type": "oz"}, "harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"}, "idle_timeout_minutes": 1, + "inference_providers": { + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, "mcp_servers": { "foo": { "args": ["string"], @@ -420,6 +452,13 @@ async def test_method_create_with_all_params(self, async_client: AsyncOzAPI) -> "warp_id": "warp_id", } }, + "memory_stores": [ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], "model_id": "model_id", "name": "name", "session_sharing": {"public_access": "VIEWER"}, @@ -530,6 +569,12 @@ async def test_method_update_with_all_params(self, async_client: AsyncOzAPI) -> "harness": {"type": "oz"}, "harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"}, "idle_timeout_minutes": 1, + "inference_providers": { + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, "mcp_servers": { "foo": { "args": ["string"], @@ -540,6 +585,13 @@ async def test_method_update_with_all_params(self, async_client: AsyncOzAPI) -> "warp_id": "warp_id", } }, + "memory_stores": [ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], "model_id": "model_id", "name": "name", "session_sharing": {"public_access": "VIEWER"}, diff --git a/tests/api_resources/test_agent.py b/tests/api_resources/test_agent.py index a10a54e..980b5b2 100644 --- a/tests/api_resources/test_agent.py +++ b/tests/api_resources/test_agent.py @@ -164,6 +164,12 @@ def test_method_run_with_all_params(self, client: OzAPI) -> None: "harness": {"type": "oz"}, "harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"}, "idle_timeout_minutes": 1, + "inference_providers": { + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, "mcp_servers": { "foo": { "args": ["string"], @@ -174,6 +180,13 @@ def test_method_run_with_all_params(self, client: OzAPI) -> None: "warp_id": "warp_id", } }, + "memory_stores": [ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], "model_id": "model_id", "name": "name", "session_sharing": {"public_access": "VIEWER"}, @@ -361,6 +374,12 @@ async def test_method_run_with_all_params(self, async_client: AsyncOzAPI) -> Non "harness": {"type": "oz"}, "harness_auth_secrets": {"claude_auth_secret_name": "claude_auth_secret_name"}, "idle_timeout_minutes": 1, + "inference_providers": { + "aws": { + "disabled": True, + "role_arn": "role_arn", + } + }, "mcp_servers": { "foo": { "args": ["string"], @@ -371,6 +390,13 @@ async def test_method_run_with_all_params(self, async_client: AsyncOzAPI) -> Non "warp_id": "warp_id", } }, + "memory_stores": [ + { + "access": "read_write", + "instructions": "instructions", + "uid": "uid", + } + ], "model_id": "model_id", "name": "name", "session_sharing": {"public_access": "VIEWER"}, diff --git a/tests/test_models.py b/tests/test_models.py index c0b225d..ae6ec09 100644 --- a/tests/test_models.py +++ b/tests/test_models.py @@ -1,7 +1,8 @@ import json -from typing import TYPE_CHECKING, Any, Dict, List, Union, Optional, cast +from typing import TYPE_CHECKING, Any, Dict, List, Union, Iterable, Optional, cast from datetime import datetime, timezone -from typing_extensions import Literal, Annotated, TypeAliasType +from collections import deque +from typing_extensions import Literal, Annotated, TypedDict, TypeAliasType import pytest import pydantic @@ -9,7 +10,7 @@ from oz_agent_sdk._utils import PropertyInfo from oz_agent_sdk._compat import PYDANTIC_V1, parse_obj, model_dump, model_json -from oz_agent_sdk._models import DISCRIMINATOR_CACHE, BaseModel, construct_type +from oz_agent_sdk._models import DISCRIMINATOR_CACHE, BaseModel, EagerIterable, construct_type class BasicModel(BaseModel): @@ -961,3 +962,56 @@ def __getattr__(self, attr: str) -> Item: ... assert model.a.prop == 1 assert isinstance(model.a, Item) assert model.other == "foo" + + +# NOTE: Workaround for Pydantic Iterable behavior. +# Iterable fields are replaced with a ValidatorIterator and may be consumed +# during serialization, which can cause subsequent dumps to return empty data. +# See: https://github.com/pydantic/pydantic/issues/9541 +@pytest.mark.parametrize( + "data, expected_validated", + [ + ([1, 2, 3], [1, 2, 3]), + ((1, 2, 3), (1, 2, 3)), + (set([1, 2, 3]), set([1, 2, 3])), + (iter([1, 2, 3]), [1, 2, 3]), + ([], []), + ((x for x in [1, 2, 3]), [1, 2, 3]), + (map(lambda x: x, [1, 2, 3]), [1, 2, 3]), + (frozenset([1, 2, 3]), frozenset([1, 2, 3])), + (deque([1, 2, 3]), deque([1, 2, 3])), + ], + ids=["list", "tuple", "set", "iterator", "empty", "generator", "map", "frozenset", "deque"], +) +@pytest.mark.skipif(PYDANTIC_V1, reason="this is only supported in pydantic v2") +def test_iterable_construction(data: Iterable[int], expected_validated: Iterable[int]) -> None: + class TypeWithIterable(TypedDict): + items: EagerIterable[int] + + class Model(BaseModel): + data: TypeWithIterable + + m = Model.model_validate({"data": {"items": data}}) + assert m.data["items"] == expected_validated + + # Verify repeated dumps don't lose data (the original bug) + assert m.model_dump()["data"]["items"] == list(expected_validated) + assert m.model_dump()["data"]["items"] == list(expected_validated) + + +@pytest.mark.skipif(PYDANTIC_V1, reason="this is only supported in pydantic v2") +def test_iterable_construction_str_falls_back_to_list() -> None: + # str is iterable (over chars), but str(list_of_chars) produces the list's repr + # rather than reconstructing a string from items. We special-case str to fall + # back to list instead of attempting reconstruction. + class TypeWithIterable(TypedDict): + items: EagerIterable[str] + + class Model(BaseModel): + data: TypeWithIterable + + m = Model.model_validate({"data": {"items": "hello"}}) + + # falls back to list of chars rather than calling str(["h", "e", "l", "l", "o"]) + assert m.data["items"] == ["h", "e", "l", "l", "o"] + assert m.model_dump()["data"]["items"] == ["h", "e", "l", "l", "o"]