Vision Agents requires a Stream account
for real-time transport. Most providers offer free tiers to get started.
Installation
LLM (Responses API)
Uses the Responses API (default for GPT-5+).| Name | Type | Default | Description |
|---|---|---|---|
model | str | — | Model (e.g., "gpt-5.4") |
api_key | str | None | API key (defaults to OPENAI_API_KEY env var) |
base_url | str | None | Custom API endpoint |
ChatCompletionsLLM
Works with any OpenAI-compatible API (Together AI, Fireworks, DeepSeek, etc.).Function Calling
Events
The OpenAI plugin emits a low-level event for raw stream data. Most developers should use the core events (LLMResponseCompletedEvent, RealtimeUserSpeechTranscriptionEvent, etc.) instead.Next Steps
OpenAI Realtime
Speech-to-speech over WebRTC
OpenAI TTS
Text-to-speech synthesis

