Skip to main content

Cartesia Voice Agents SDK

Project description

Cartesia Line SDK

Build intelligent, low-latency voice agents with Line.

Line brings voice to your text agents with Cartesia's state-of-the-art speech models. We handle audio orchestration, deployment, and observability so you can focus on your agent's reasoning.

Features

  • Real-time interruption support — Handles audio interruptions and turn-taking out-of-the-box
  • Tool calling — Connect to databases, APIs, and external services
  • Multi-agent handoffs — Route conversations between specialized agents
  • Web search — Built-in tool for real-time information lookup
  • 100+ LLM providers — Works with any LLM via LiteLLM
  • Instant deployment — Build, deploy, and start talking in minutes

Quick Start

1. Clone and run an example:

git clone https://github.com/cartesia-ai/line.git
cd line/examples/basic_chat
GEMINI_API_KEY=your-key uv run python main.py

2. Or create from scratch:

mkdir my-agent && cd my-agent
uv init && uv add cartesia-line

Create main.py:

import os
from line.llm_agent import LlmAgent, LlmConfig, end_call
from line.voice_agent_app import VoiceAgentApp

async def get_agent(env, call_request):
    return LlmAgent(
        model="gemini/gemini-2.5-flash-preview-09-2025",
        api_key=os.getenv("GEMINI_API_KEY"),
        tools=[end_call],
        config=LlmConfig(
            system_prompt="You are a helpful voice assistant.",
            introduction="Hello! How can I help you today?",
        ),
    )

app = VoiceAgentApp(get_agent=get_agent)

if __name__ == "__main__":
    app.run()

Run it:

GEMINI_API_KEY=your-key uv run python main.py

3. (Optional) Install the CLI to test locally:

curl -fsSL https://cartesia.sh | sh

Then chat with your agent:

PORT=8000 uv run python main.py
cartesia chat 8000

See the CLI documentation for deployment and management commands.


Customize Your Agent's Prompt

System Prompt & Introduction

Configure your agent's personality and behavior via LlmConfig:

config = LlmConfig(
    system_prompt="You are a customer service agent for Acme Corp. Be friendly and concise.",
    introduction="Hi! Thanks for calling Acme. How can I help?",
)
  • system_prompt — Defines the agent's personality, rules, and context
  • introduction — First message spoken when the call starts (set to "" to wait for user)

Dynamic Prompts from API

Use LlmConfig.from_call_request() to configure prompts dynamically from your API:

async def get_agent(env: AgentEnv, call_request: CallRequest):
    # Prompts come from call_request.agent.system_prompt and call_request.agent.introduction
    # Falls back to your defaults if not provided
    return LlmAgent(
        model="gemini/gemini-2.5-flash-preview-09-2025",
        tools=[end_call],
        config=LlmConfig.from_call_request(
            call_request,
            fallback_system_prompt="You are a helpful assistant.",
            fallback_introduction="Hello! How can I help?",
        ),
    )

Add Tools to Your Agent

Built-in Tools

Ready-to-use tools for common actions:

from line.llm_agent import LlmAgent, LlmConfig, end_call, send_dtmf, transfer_call, web_search

agent = LlmAgent(
    model="gemini/gemini-2.5-flash-preview-09-2025",
    tools=[end_call, send_dtmf, transfer_call, web_search],
    config=LlmConfig(...),
)
Tool What it does
end_call Ends the call
send_dtmf Presses phone buttons (0-9, *, #)
transfer_call Transfers to a phone number (E.164 format)
web_search Searches the web (native LLM search or DuckDuckGo fallback)

Loopback Tools — Fetch Data & Call APIs

Results go back to the LLM for a natural language response:

from typing import Annotated
from line.llm_agent import loopback_tool

@loopback_tool
async def get_order_status(ctx, order_id: Annotated[str, "The order ID"]) -> str:
    """Look up order status."""
    order = await db.get_order(order_id)
    return f"Order {order_id}: {order.status}"

agent = LlmAgent(tools=[get_order_status, end_call], ...)

User: "What's the status of order 12345?" Agent: Calls tool → LLM responds: "Your order was delivered on January 5th!"

Passthrough Tools — Deterministic Actions

Output goes directly to the user, bypassing the LLM:

from typing import Annotated
from line.events import AgentSendText, AgentTransferCall
from line.llm_agent import passthrough_tool

@passthrough_tool
async def transfer_to_support(ctx, reason: Annotated[str, "Why they need support"]):
    """Transfer to support team."""
    yield AgentSendText(text="Transferring you to support now.")
    yield AgentTransferCall(target_phone_number="+18005551234")

agent = LlmAgent(tools=[transfer_to_support, end_call], ...)

Handoff Tools — Multi-Agent Workflows

Transfer control to a specialized agent:

from line.llm_agent import LlmAgent, LlmConfig, agent_as_handoff, end_call

spanish_agent = LlmAgent(
    model="anthropic/claude-sonnet-4-5",
    tools=[end_call],
    config=LlmConfig(
        system_prompt="You speak only in Spanish.",
        introduction="¡Hola! ¿Cómo puedo ayudarte?",
    ),
)

main_agent = LlmAgent(
    model="gemini/gemini-2.5-flash-preview-09-2025",
    tools=[
        end_call,
        agent_as_handoff(
            spanish_agent,
            handoff_message="Transferring you to our Spanish-speaking agent...",
            name="transfer_to_spanish",
            description="Transfer when user wants to speak Spanish.",
        ),
    ],
    config=LlmConfig(system_prompt="Transfer to Spanish if requested."),
)

Tool Types Summary

Type How to create Result goes to Use for
Loopback @loopback_tool Back to LLM API calls, data lookup
Passthrough @passthrough_tool Directly to user Deterministic actions
Handoff agent_as_handoff() or @handoff_tool Another agent Multi-agent workflows

Long-Running Tools

By default, tool calls are terminated when the agent is interrupted (though any reasoning and tool call response values already produced are preserved for use in the next generation).

For tools that take a long time to complete, set is_background=True. The tool will continue running in the background until completion regardless of interruptions, then loop back to the LLM:

from typing import Annotated
from line.llm_agent import loopback_tool

@loopback_tool(is_background=True)
async def search_database(ctx, query: Annotated[str, "Search query"]) -> str:
    """Search that may take a while."""
    results = await slow_database_search(query)
    return results

Customize Your Agent's Implementation

Wrap with Custom Logic

Implement the Agent protocol to add guardrails, logging, or preprocessing:

from line.agent import TurnEnv
from line.events import InputEvent, OutputEvent, UserTurnEnded, AgentSendText
from line.llm_agent import LlmAgent, LlmConfig, end_call

class GuardedAgent:
    def __init__(self, inner_agent):
        self.inner = inner_agent
        self.blocked_words = ["competitor", "confidential"]

    async def process(self, env: TurnEnv, event: InputEvent):
        # Pre-process: check user input for blocked words
        if isinstance(event, UserTurnEnded):
            user_text = " ".join(
                item.content for item in event.content if hasattr(item, "content")
            )
            if any(word in user_text.lower() for word in self.blocked_words):
                yield AgentSendText(text="I can't discuss that topic.")
                return

        # Delegate to inner agent
        async for output in self.inner.process(env, event):
            yield output

async def get_agent(env, call_request):
    inner = LlmAgent(
        model="gemini/gemini-2.5-flash-preview-09-2025",
        tools=[end_call],
        config=LlmConfig(system_prompt="You are a helpful assistant."),
    )
    return GuardedAgent(inner)

LLM Provider Support

Line leverages LiteLLM to support 100+ LLM providers. Pass any LiteLLM-compatible model string to LlmAgent:

Provider Model format
OpenAI gpt-5-nano, gpt-5.2
Anthropic anthropic/claude-haiku-4-5-20251001, anthropic/claude-sonnet-4-5
Google gemini/gemini-2.5-flash-preview-09-2025, gemini/gemini-3.0-preview

Agent Examples

Example Description
Basic Chat Simple conversational agent
Form Filler Collect structured data
Phone Transfer IVR navigation & transfers
Multi-Agent Hand off between agents
Echo Tool Custom handoff tool

Integrations

Integration Description
Exa Web Research Real-time web search
Browserbase Fill web forms via voice

Documentation

Getting Help

Acknowledgments

Line leverages the fantastic work by the maintainers of LiteLLM. Their open-source library provides the unified LLM interface that makes it possible to support 100+ providers out of the box.

LiteLLM is licensed under the MIT License.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

cartesia_line-0.2.1.tar.gz (63.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

cartesia_line-0.2.1-py3-none-any.whl (50.1 kB view details)

Uploaded Python 3

File details

Details for the file cartesia_line-0.2.1.tar.gz.

File metadata

  • Download URL: cartesia_line-0.2.1.tar.gz
  • Upload date:
  • Size: 63.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for cartesia_line-0.2.1.tar.gz
Algorithm Hash digest
SHA256 243f3e31cfabbda7c236a95ef7ac7c2304f4e7eeb76f4efcda769034639bd247
MD5 e694c841c1053b4e326108bbe3389763
BLAKE2b-256 fa281e4fa256a9a0ab20daf4e731a25378be3423b5f21a00c7d0fcd106989abf

See more details on using hashes here.

Provenance

The following attestation bundles were made for cartesia_line-0.2.1.tar.gz:

Publisher: publish-to-pypi.yaml on cartesia-ai/line

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file cartesia_line-0.2.1-py3-none-any.whl.

File metadata

  • Download URL: cartesia_line-0.2.1-py3-none-any.whl
  • Upload date:
  • Size: 50.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for cartesia_line-0.2.1-py3-none-any.whl
Algorithm Hash digest
SHA256 737acfdf37591313b7ac1ff6e165d51eaf0e3561b4227190ee92d5f8ea65f6a9
MD5 66aa696a96a1552787942689ba495d35
BLAKE2b-256 58bbfe0898412f3cb7cdf49587e27d7a5c07f7f24c0b0a5eb8eb264d38a8abdb

See more details on using hashes here.

Provenance

The following attestation bundles were made for cartesia_line-0.2.1-py3-none-any.whl:

Publisher: publish-to-pypi.yaml on cartesia-ai/line

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page