The Gemini Live API enables low-latency bidirectional text and voice interactions with Gemini. Using the Live API, you can provide end users with the experience of natural, human-like voice conversations, with the ability to interrupt the model's responses using text or voice commands. The model can process text and audio input (video coming soon!), and it can provide text and audio output.
You can prototype with prompts and the Live API in Google AI Studio or Vertex AI Studio.
The Live API is a stateful API that creates a WebSocket connection to establish a session between the client and the Gemini server. For details, see the Live API reference documentation (Gemini Developer API | Vertex AI Gemini API).
Before you begin
Click your Gemini API provider to view provider-specific content and code on this page. |
If you haven't already, complete the
getting started guide,
which describes how to set up your Firebase project,
connect your app to Firebase, add the SDK,
initialize the backend service for your chosen Gemini API provider, and
create a LiveModel
instance.