Voice Agent API
New Features
🤖 GPT 5.0 LLM Support
We’ve added support for OpenAI’s new 5.0 models in our Voice Agent API!
Available Models:
- 5.0 (gpt-5)
- 5.0 Mini (gpt-5-mini)
- 5.0 Nano (gpt-5-nano)
Implementation: Configure your chosen model in your Voice Agent settings:
For complete information about supported LLMs including the GPT 5.0 models, visit our Voice Agent LLM Models documentation.
Voice Agent API
New Features
🤖 GPT-OSS-20B LLM Support
We’ve added support for OpenAI’s first open source LLM, gpt-oss-20b, in our Voice Agent API!
Available Models:
- GPT OSS 20B
Implementation: Configure GPT-OSS-20B in your Voice Agent settings:
For complete information about supported LLMs including GPT OSS 20B, visit our Voice Agent LLM Models documentation.
Voice Agent API
New Features
🎯 Smart Formatting for More Readable Conversations
We’ve added a new smart_format
option to improve transcript readability in UI applications. This feature enables smart formatting for better user experience when displaying transcribed conversation.
Key Features:
- Enhanced transcript formatting for UI applications
- Defaults to
false
for backward compatibility
Implementation:
Configure the smart_format
option in your Voice Agent listen provider settings:
For complete implementation details, see our Voice Agent configuration documentation.
🔒 Model Improvement Program Opt-Out
Users can now opt out of our Model Improvement Program when using the Voice Agent API.
Implementation:
Add mip_opt_out: true
to your Settings message:
For more information about the Model Improvement Program and opt-out options, visit our Model Improvement Partnership Program documentation.
🤖 Gemini LLM Support
We’ve added support for Google’s Gemini LLMs in our Voice Agent API! This expands our LLM options to include Google’s powerful language models.
Available Models:
- Gemini 2.5 Flash
- Gemini 2.0 Flash
- Gemini 2.0 Flash Lite
For complete information about supported LLMs including Gemini models, visit our Voice Agent LLM Models documentation.
Voice Agent API
New Features
🤖 Expanded LLM Support
We’ve significantly expanded our LLM options across our two pricing tiers. The following LLMs are now available for use:
Standard Models
- OpenAI GPT-4.1 mini
- OpenAI GPT-4.1 nano
- OpenAI GPT-4o mini
- Anthropic Claude Haiku 3.5
Advanced Models
- OpenAI GPT-4.1
- OpenAI GPT-4o
- Anthropic Claude Sonnet 4
For complete information about supported LLMs, visit our Voice Agent LLM Models documentation or try them out in our API Playground.
🌍 Spanish Language Support
Voice Agents now support Spanish conversations with the addition of Aura-2 Spanish TTS. Configure your agent’s language settings to enable Spanish voice interactions.
See our Voice Agent API documentation for implementation details.
💬 Conversation Context Feature
Introducing comprehensive conversation continuity with our new context feature:
Complete Context Awareness
- Agents maintain conversation history across sessions
- Seamless continuation of previous interactions
Enhanced User Experience
- More natural conversations with historical context
- Consistent interaction patterns across sessions
Flexible Implementation
- Support for both conversational and function call history
- Configurable history settings
Implementation
Use the agent.context
object to provide conversation history when starting new sessions:
Conversational Messages:
Function Call Messages:
To disable function call history, set settings.flags.history
to false
in the Settings message.
Documentation
🔍 Enhanced Error Visibility
We’ve improved client-side visibility of LLM and TTS errors to provide better debugging and user experience.