Drop-in interrupt-resume for the google-genai Live API. One function call, zero code changes.
Project description
gemini-live-interrupt
Automatic interrupt-resume for the Gemini Live API. When a user interrupts the model mid-sentence, this package injects context behind the scenes so the model knows what it was saying, what the user said, and can decide whether to continue, pivot, or blend both.
Works with any existing google-genai application. No code changes required beyond a single function call.
Install
pip install gemini-live-interrupt
The Problem
When using the Gemini Live API for real-time audio conversations, users can interrupt the model at any time. The API sends an interrupted signal, but the model has no memory of what it was saying or how far it got. On its next response, it starts fresh — losing the context of the interrupted response entirely.
For example, if the model is telling a 5-line story and the user interrupts at line 2, the model won't know to continue from line 2. It either starts over, or moves on as if the story never happened.
How This Package Solves It
Call enable_interrupt_resume() once before you connect to the Gemini Live API. After that, every live session automatically:
- Tracks what the model was saying (output transcription)
- Tracks what the user said (input transcription)
- Detects when an interruption happens
- Injects a context-recovery prompt to the model with what was said and what to do next
The model then decides the right course of action based on the context.
Step-by-step Usage
Step 1: Install the package
pip install gemini-live-interrupt
Step 2: Add one line to your existing code
Add enable_interrupt_resume() before you create any live sessions:
from gemini_live_interrupt import enable_interrupt_resume
enable_interrupt_resume()
That's it. Your existing google-genai code does not need any other changes.
Step 3: Make sure transcription is enabled in your config
The package reads the transcription text that the Gemini Live API provides. For this to work, your LiveConnectConfig must have input and output transcription enabled:
config = {
"response_modalities": ["AUDIO"],
"input_audio_transcription": {},
"output_audio_transcription": {},
# ... your other config
}
If transcription is not enabled, the package has nothing to track and interruptions won't be handled.
Step 4: Connect and use the API as normal
from google import genai
from google.genai import types
from gemini_live_interrupt import enable_interrupt_resume
# Step 2: enable before connecting
enable_interrupt_resume()
# Your existing code — nothing changes below this line
client = genai.Client(vertexai=True, project="my-project", location="us-central1")
config = {
"response_modalities": [types.Modality.AUDIO],
"input_audio_transcription": {},
"output_audio_transcription": {},
"speech_config": {
"voice_config": {
"prebuilt_voice_config": {"voice_name": "Aoede"}
}
},
}
async with client.aio.live.connect(
model="gemini-live-2.5-flash-native-audio", config=config
) as session:
# send audio, receive responses — interruptions are handled automatically
async for msg in session.receive():
if msg.server_content and msg.server_content.model_turn:
for part in msg.server_content.model_turn.parts:
if part.inline_data:
# play or forward audio as usual
play_audio(part.inline_data.data)
What Happens When the User Interrupts
When the model is speaking and the user starts talking, the API sends content.interrupted. This package intercepts it and sends a context prompt to the model containing:
- Recent conversation — the last few turns of dialogue so the model has context
- What was being said — the output transcription accumulated before the interruption
- What the user said — the input transcription of the interrupting speech
- Exact continuation text — the sentence that was cut off, from its beginning (not mid-word)
- Instructions for how to proceed:
- New question: answer it, then ask if the user wants to continue the interrupted response
- Background noise / acknowledgment: seamlessly continue from the sentence that was cut off
- Follow-up or clarification: address it briefly, then continue the interrupted response
The model receives all of this before generating its next response, so it can make an informed decision.
Example: FastAPI WebSocket App
from fastapi import FastAPI, WebSocket
from google import genai
from google.genai import types
from gemini_live_interrupt import enable_interrupt_resume
app = FastAPI()
enable_interrupt_resume()
client = genai.Client(vertexai=True, project="my-project", location="us-central1")
@app.websocket("/live")
async def live_audio(ws: WebSocket):
await ws.accept()
config = {
"response_modalities": [types.Modality.AUDIO],
"input_audio_transcription": {},
"output_audio_transcription": {},
}
async with client.aio.live.connect(
model="gemini-live-2.5-flash-native-audio", config=config
) as session:
async for msg in session.receive():
if msg.server_content and msg.server_content.model_turn:
for part in msg.server_content.model_turn.parts:
if part.inline_data:
await ws.send_bytes(part.inline_data.data)
Interruptions are handled transparently. When a connected client interrupts the model, the resume context is injected automatically.
Custom Prompt
If you want to control exactly what prompt is injected on interruption, pass a prompt_builder function:
def my_prompt(heard: str, user_text: str, history: list) -> str:
"""
Args:
heard: what the model was saying (output transcription up to interruption)
user_text: what the user said to interrupt
history: list of (role, text) tuples — recent conversation turns
Returns:
the prompt string to inject
"""
return (
f"You were interrupted. You were saying: {heard}. "
f"The user said: {user_text}. "
f"Continue from where you left off."
)
enable_interrupt_resume(prompt_builder=my_prompt)
Configuration
| Parameter | Type | Default | Description |
|---|---|---|---|
prompt_builder |
callable or None |
None |
Custom (heard, user_text, history) -> str function. If None, uses the built-in prompt. |
max_history |
int |
4 |
Number of conversation turns to keep (4 = last 2 exchanges). |
Disabling
To remove the patch and restore original behavior:
from gemini_live_interrupt import disable_interrupt_resume
disable_interrupt_resume()
Requirements
- Python >= 3.11
google-genaiSDKwrapt
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file gemini_live_interrupt-0.1.0.tar.gz.
File metadata
- Download URL: gemini_live_interrupt-0.1.0.tar.gz
- Upload date:
- Size: 7.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.6.14
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
80e0df34329def0333c377a0b4f75e4e7c6a35854519c623bfafac1d5103c7a8
|
|
| MD5 |
fc45687c1a4d2d2c539cbd4e9ca4a9c1
|
|
| BLAKE2b-256 |
2026bda5dd5b8cc6bdc2a321da860a8120aa62103a6e738928165d187df4d005
|
File details
Details for the file gemini_live_interrupt-0.1.0-py3-none-any.whl.
File metadata
- Download URL: gemini_live_interrupt-0.1.0-py3-none-any.whl
- Upload date:
- Size: 7.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.6.14
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
a24bc558f710e4041053ecedde99d00c2c76f2d7037ca198926769af65ed4d8f
|
|
| MD5 |
2f86258d905518b510ed3dace73f591d
|
|
| BLAKE2b-256 |
0fca3a38ccda5d19f35ea3de5fae26b2a98ddd5a5c85ca781938e8120e853cba
|