@@ -12,7 +12,7 @@ This example shows you how to build a basic video AI agent using [Vision Agents]
1212- Python 3.13 or higher
1313- API keys for:
1414 - [ OpenAI] ( https://openai.com ) (for the LLM)
15- - [ Cartesia ] ( https://cartesia.ai / ) (for text-to-speech)
15+ - [ Elevenlabs ] ( https://elevenlabs.io / ) (for text-to-speech)
1616 - [ Deepgram] ( https://deepgram.com/ ) (for speech-to-text)
1717 - [ Stream] ( https://getstream.io/ ) (for video/audio infrastructure)
1818 - [ Smart Turn] ( https://fal.ai/models/fal-ai/smart-turn ) (for turn detection)
@@ -27,11 +27,10 @@ This example shows you how to build a basic video AI agent using [Vision Agents]
27272 . Create a ` .env ` file with your API keys:
2828 ```
2929 OPENAI_API_KEY=your_openai_key
30- CARTESIA_API_KEY=your_cartesia_key
30+ ELEVENLABS_API_KEY=your_11labs_key
3131 DEEPGRAM_API_KEY=your_deepgram_key
3232 STREAM_API_KEY=your_stream_key
3333 STREAM_API_SECRET=your_stream_secret
34- FAL_KEY=your_fal_key
3534 ```
3635
3736## Running the Example
@@ -59,7 +58,7 @@ agent = Agent(
5958 agent_user = User(name = " My happy AI friend" , id = " agent" ),
6059 instructions = " You're a video AI assistant..." ,
6160 llm = openai.LLM(model = " gpt-4o-mini" ),
62- tts = cartesia .TTS(),
61+ tts = elevenlabs .TTS(),
6362 stt = deepgram.STT(),
6463 turn_detection = smart_turn.TurnDetection(),
6564)
@@ -133,7 +132,7 @@ Edit the `instructions` parameter to change how your agent behaves. The instruct
133132You can swap out any component:
134133- Try ` openai.Realtime() ` for lower latency
135134- Use ` gemini.Realtime() ` for Google's model
136- - Switch TTS providers to ` cartesia .TTS()` or ` kokoro.TTS() `
135+ - Switch TTS providers to ` elevenlabs .TTS()` or ` kokoro.TTS() `
137136
138137### Add Processors
139138
0 commit comments