diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 7cb1a16..1ae2526 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "0.2.18-alpha.2" + ".": "0.3.0-alpha.1" } diff --git a/.stats.yml b/.stats.yml index 1f5340b..c93a005 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 106 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-7c002d994b96113926e24a0f99ff80a52b937481e383b584496087ecdc2d92d6.yml -openapi_spec_hash: e9c825e9199979fc5f754426a1334499 +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-1c5bc84a8d003bb986b2cc2a7bb6a5232ab0514469a7f67ddbf58c06de248a03.yml +openapi_spec_hash: ad377dd848973abb656ca35f5fdc93df config_hash: e67fd054e95c1e82f78f4b834e96bb65 diff --git a/CHANGELOG.md b/CHANGELOG.md index 7070ef6..a75b956 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,13 @@ # Changelog +## 0.3.0-alpha.1 (2025-08-13) + +Full Changelog: [v0.2.18-alpha.2...v0.3.0-alpha.1](https://github.com/llamastack/llama-stack-client-python/compare/v0.2.18-alpha.2...v0.3.0-alpha.1) + +### Features + +* **api:** update via SDK Studio ([1f8b267](https://github.com/llamastack/llama-stack-client-python/commit/1f8b2677361ad2a5c75f071d99622754d0fa34a7)) + ## 0.2.18-alpha.2 (2025-08-12) Full Changelog: [v0.2.18-alpha.1...v0.2.18-alpha.2](https://github.com/llamastack/llama-stack-client-python/compare/v0.2.18-alpha.1...v0.2.18-alpha.2) diff --git a/pyproject.toml b/pyproject.toml index c25d53e..8e3a66e 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "llama_stack_client" -version = "0.2.18-alpha.2" +version = "0.3.0-alpha.1" description = "The official Python library for the llama-stack-client API" dynamic = ["readme"] license = "MIT" diff --git a/src/llama_stack_client/types/create_response.py b/src/llama_stack_client/types/create_response.py index b0eaf3e..fbb519f 100644 --- a/src/llama_stack_client/types/create_response.py +++ b/src/llama_stack_client/types/create_response.py @@ -22,13 +22,7 @@ class Result(BaseModel): """ category_scores: Optional[Dict[str, float]] = None - """A list of the categories along with their scores as predicted by model. - - Required set of categories that need to be in response - violence - - violence/graphic - harassment - harassment/threatening - hate - - hate/threatening - illicit - illicit/violent - sexual - sexual/minors - - self-harm - self-harm/intent - self-harm/instructions - """ + """A list of the categories along with their scores as predicted by model.""" user_message: Optional[str] = None