Multilingual Voice Agents
Learn the best way to support a multi-lingual voice agent through Deepgram.
Select a multilingual speech-to-text (STT) model.
In order to have a fully multi-lingual agent, you need to select a multi-lingual speech-to-text model and specify the appropriate language. Today, you should specify the following parameters in your Settings.
agent.listen.provider.model:nova-3agent.listen.provider.language:multi
Select a multilingual text-to-speech (TTS) provider.
OpenAI, Eleven Labs, and Cartesia have multi-lingual language models. Select one of those providers and use agent.speak.provider.language as multi. For Eleven Labs, this parameter aligns to their language_code.
Bilingual English/Spanish with Deepgram TTS
You can also build a bilingual English/Spanish voice agent using Deepgram’s own Aura text-to-speech. Deepgram offers Spanish voices that seamlessly switch between English and Spanish, known as codeswitching voices: Aquila, Carina, Diana, Javier, and Selena.
To use a codeswitching voice, set your TTS provider to deepgram and specify one of the codeswitching voice models:
agent.speak.provider.type:deepgramagent.speak.provider.model:aura-2-aquila-es(oraura-2-carina-es,aura-2-diana-es,aura-2-javier-es,aura-2-selena-es)
These voices handle mixed-language responses naturally, so the agent can respond in both English and Spanish within the same conversation without switching TTS providers.
For the full list of Spanish voices and details, see Voices and Languages.
Prompt recommendations
Prompt design can help adjust the agent’s behavior depending on your expected use case. Results may vary depending on the LLM provider you use.