AI-powered podcast generation tool that creates conversational audio content from text sources
Project description
Podcast Creator
An AI-powered podcast generation library that creates conversational audio content from text-based sources. This pip-installable package processes documents, generates structured outlines, creates natural dialogue transcripts, and converts them into high-quality audio podcasts using LangGraph workflow orchestration.
๐ง Live Demo
Listen to a real podcast generated with this tool - a 4-person debate on the Situational Awareness Paper. Includes my own cloned voice ๐
Generated using the diverse_panel episode profile with 4 AI experts discussing the nuances of artificial general intelligence and situational awareness.
And here is a one-speaker version of it, like your real dedicated teacher.
๐ Quick Start
Installation
# Library only (programmatic use)
uv add podcast-creator
# or pip install podcast-creator
# Full installation with web UI
uv add podcast-creator --extra ui
# or pip install podcast-creator[ui]
# Or install from source
git clone <repository-url>
cd podcast-creator
uv sync
# Don't have uv? Install it with:
# curl -LsSf https://astral.sh/uv/install.sh | sh
# or
# pip install uv
Installation Options:
- Library only:
pip install podcast-creator- For programmatic use without web interface - With UI:
pip install podcast-creator[ui]- Includes Streamlit web interface for visual management
Configure API Keys
# Copy the example environment file
cp .env.example .env
# Edit .env and add your API keys:
# - OpenAI API key for LLM models
# - ElevenLabs API key for high-quality TTS
# - Other provider keys as needed
Initialize Your Project
# Create templates and configuration files
podcast-creator init
# This creates:
# - prompts/podcast/outline.jinja
# - prompts/podcast/transcript.jinja
# - speakers_config.json
# - episodes_config.json
# - example_usage.py
Generate Your First Podcast
๐จ New: Web Interface
# Launch the Streamlit web interface
podcast-creator ui
# Custom port/host
podcast-creator ui --port 8080 --host 0.0.0.0
# The UI provides:
# - Visual profile management
# - Multi-content podcast generation
# - Episode library with playback
# - Import/export functionality
๐ Episode Profiles (Streamlined)
import asyncio
from podcast_creator import create_podcast
async def main():
# One-liner podcast creation with episode profiles!
result = await create_podcast(
content="Your content here...",
episode_profile="tech_discussion", # ๐ฏ Pre-configured settings
episode_name="my_podcast",
output_dir="output/my_podcast"
)
print(f"โ
Podcast created: {result['final_output_file_path']}")
asyncio.run(main())
๐ Classic: Full Configuration
import asyncio
from podcast_creator import create_podcast
async def main():
result = await create_podcast(
content="Your content here...",
briefing="Create an engaging discussion about...",
episode_name="my_podcast",
output_dir="output/my_podcast",
speaker_config="ai_researchers"
)
print(f"โ
Podcast created: {result['final_output_file_path']}")
asyncio.run(main())
๐ฏ Episode Profiles - Streamlined Podcast Creation
Episode Profiles are pre-configured sets of podcast generation parameters that enable one-liner podcast creation for common use cases while maintaining full customization flexibility.
๐ Why Episode Profiles?
- 67% fewer parameters to specify for common use cases
- Consistent configurations across podcast series
- Faster iteration and prototyping
- Team collaboration with shared settings
- Full backward compatibility with existing code
๐ Bundled Profiles
| Profile | Description | Speakers | Segments | Use Case |
|---|---|---|---|---|
tech_discussion |
Technology topics with expert analysis | 2 AI researchers | 4 | Technical content, AI/ML topics |
solo_expert |
Educational explanations | 1 expert teacher | 3 | Learning content, tutorials |
business_analysis |
Market and business insights | 3 business analysts | 4 | Business strategy, market analysis |
diverse_panel |
Multi-perspective discussions | 4 diverse voices | 5 | Complex topics, debate-style content |
๐ช Usage Patterns
# 1. Simple profile usage
result = await create_podcast(
content="Your content...",
episode_profile="tech_discussion",
episode_name="my_podcast",
output_dir="output/my_podcast"
)
# 2. Profile with briefing suffix
result = await create_podcast(
content="Your content...",
episode_profile="business_analysis",
briefing_suffix="Focus on ROI and cost optimization",
episode_name="my_podcast",
output_dir="output/my_podcast"
)
# 3. Profile with parameter overrides
result = await create_podcast(
content="Your content...",
episode_profile="solo_expert",
outline_model="gpt-4o", # Override default
num_segments=5, # Override default
episode_name="my_podcast",
output_dir="output/my_podcast"
)
๐ง Custom Episode Profiles
from podcast_creator import configure
# Define your own episode profiles
configure("episode_config", {
"profiles": {
"my_startup_pitch": {
"speaker_config": "business_analysts",
"outline_model": "gpt-4o",
"default_briefing": "Create an engaging startup pitch...",
"num_segments": 6
}
}
})
# Use your custom profile
result = await create_podcast(
content="Your content...",
episode_profile="my_startup_pitch",
episode_name="pitch_deck",
output_dir="output/pitch_deck"
)
โจ Features
๐ง Flexible Configuration
from podcast_creator import configure
# Configure with custom templates
configure("templates", {
"outline": "Your custom outline template...",
"transcript": "Your custom transcript template..."
})
# Configure with custom paths
configure({
"prompts_dir": "./my_templates",
"speakers_config": "./my_speakers.json",
"output_dir": "./podcasts"
})
# Configure speakers inline
configure("speakers_config", {
"profiles": {
"my_hosts": {
"tts_provider": "elevenlabs",
"tts_model": "eleven_flash_v2_5",
"speakers": [...]
}
}
})
๐๏ธ Core Features
- ๐จ Web Interface: Complete Streamlit UI for visual podcast creation
- ๐ฏ Episode Profiles: Pre-configured settings for one-liner podcast creation
- ๐ LangGraph Workflow: Advanced state management and parallel processing
- ๐ Automatic Retry: Exponential backoff for transient API failures (LLM & TTS)
- ๐ฅ Multi-Speaker Support: Dynamic 1-4 speaker configurations with rich personalities
- โก Parallel Audio Generation: API-safe batching with concurrent processing
- ๐ง Fully Configurable: Multiple AI providers (OpenAI, Anthropic, Google, etc.)
- ๐ Multi-Content Support: Combine text, files, and URLs in structured arrays
- ๐ค AI-Powered Generation: Creates structured outlines and natural dialogues
- ๐ต Multi-Provider TTS: ElevenLabs, OpenAI, Google TTS support
- ๐ Flexible Templates: Jinja2-based prompt customization
- ๐ Multilingual Support: Generate content in multiple languages
- ๐ Episode Library: Built-in audio playback and transcript viewing
๐๏ธ Architecture
Configuration Priority
The library uses a smart priority system for loading resources:
-
User Configuration (highest priority)
configure("templates", {"outline": "...", "transcript": "..."})
-
Custom Paths
configure("prompts_dir", "/path/to/templates")
-
Working Directory
./prompts/podcast/*.jinja./speakers_config.json./episodes_config.json
-
Bundled Defaults (lowest priority)
- Package includes production-ready templates
- Multiple speaker profiles included
๐ Usage Examples
๐ฏ Episode Profiles (Recommended)
import asyncio
from podcast_creator import create_podcast
# Simple episode profile usage
async def main():
result = await create_podcast(
content="AI has transformed many industries...",
episode_profile="tech_discussion", # ๐ One-liner magic!
episode_name="ai_impact",
output_dir="output/ai_impact"
)
asyncio.run(main())
๐ Classic Configuration
import asyncio
from podcast_creator import create_podcast
async def main():
result = await create_podcast(
content="AI has transformed many industries...",
briefing="Create an informative discussion about AI impact",
episode_name="ai_impact",
output_dir="output/ai_impact",
speaker_config="ai_researchers"
)
asyncio.run(main())
Advanced Configuration
from podcast_creator import configure, create_podcast
# Custom speaker configuration (with optional per-speaker TTS overrides)
configure("speakers_config", {
"profiles": {
"tech_experts": {
"tts_provider": "elevenlabs",
"tts_model": "eleven_flash_v2_5",
"speakers": [
{
"name": "Dr. Alex Chen",
"voice_id": "your_voice_id",
"backstory": "Senior AI researcher with focus on ethics",
"personality": "Thoughtful, asks probing questions"
},
{
"name": "Jamie Rodriguez",
"voice_id": "alloy",
"backstory": "Tech journalist and startup advisor",
"personality": "Enthusiastic, great at explanations",
"tts_provider": "openai",
"tts_model": "tts-1"
}
]
}
}
})
# Custom templates
configure("templates", {
"outline": """
Create a {{ num_segments }}-part podcast outline about: {{ briefing }}
Content: {{ context }}
Speakers: {% for speaker in speakers %}{{ speaker.name }}: {{ speaker.personality }}{% endfor %}
""",
"transcript": """
Generate natural dialogue for: {{ segment.name }}
Keep it conversational and engaging.
"""
})
# Generate podcast with custom configuration
result = await create_podcast(
content="Your content...",
briefing="Your briefing...",
episode_name="custom_podcast",
speaker_config="tech_experts"
)
๐ช Episode Profile Variations
# Solo expert explanation
result = await create_podcast(
content="Technical content...",
episode_profile="solo_expert",
episode_name="deep_dive",
output_dir="output/deep_dive"
)
# Business analysis
result = await create_podcast(
content="Market trends...",
episode_profile="business_analysis",
episode_name="market_analysis",
output_dir="output/market_analysis"
)
# Panel discussion with diverse perspectives
result = await create_podcast(
content="Complex topic...",
episode_profile="diverse_panel",
episode_name="panel_discussion",
output_dir="output/panel_discussion"
)
๐ง Episode Profile Customization
# Use profile with briefing suffix
result = await create_podcast(
content="Cloud computing trends...",
episode_profile="business_analysis",
briefing_suffix="Focus on cost optimization and ROI metrics",
episode_name="cloud_economics",
output_dir="output/cloud_economics"
)
# Override specific parameters
result = await create_podcast(
content="Quantum computing...",
episode_profile="tech_discussion",
outline_model="gpt-4o", # Override default
num_segments=6, # Override default
episode_name="quantum_deep",
output_dir="output/quantum_deep"
)
๐ง Configuration API
Main Functions
from podcast_creator import configure, get_config, create_podcast
# Set configuration
configure(key, value)
configure({"key1": "value1", "key2": "value2"})
# Get configuration
value = get_config("key", default_value)
# Generate podcast
result = await create_podcast(...)
Configuration Options
| Key | Type | Description |
|---|---|---|
prompts_dir |
str |
Directory containing template files |
templates |
dict |
Inline template content |
speakers_config |
str/dict |
Path to speaker JSON or inline config |
episode_config |
str/dict |
Path to episode JSON or inline config |
output_dir |
str |
Default output directory |
๐ญ Speaker Configuration
Speaker Profile Structure
{
"profiles": {
"profile_name": {
"tts_provider": "elevenlabs",
"tts_model": "eleven_flash_v2_5",
"speakers": [
{
"name": "Speaker Name",
"voice_id": "voice_id_from_provider",
"backstory": "Rich background that informs expertise",
"personality": "Speaking style and traits"
}
]
}
}
}
Per-Speaker TTS Overrides
Individual speakers can override the profile-level TTS provider, model, and config. This lets you mix different TTS services within the same podcast โ for example, one speaker on ElevenLabs and another on OpenAI TTS.
{
"profiles": {
"mixed_providers": {
"tts_provider": "openai",
"tts_model": "tts-1",
"speakers": [
{
"name": "Dr. Sarah Chen",
"voice_id": "custom_eleven_voice_id",
"backstory": "AI researcher...",
"personality": "Analytical and methodical",
"tts_provider": "elevenlabs",
"tts_model": "eleven_flash_v2_5",
"tts_config": { "voice_settings": { "stability": 0.8 } }
},
{
"name": "Marcus Rivera",
"voice_id": "alloy",
"backstory": "Tech journalist...",
"personality": "Engaging and curious"
}
]
}
}
}
In this example, Dr. Sarah Chen uses ElevenLabs while Marcus Rivera uses the profile-level OpenAI TTS. All three fields (tts_provider, tts_model, tts_config) are optional per speaker โ any field not set falls back to the profile-level value. If a speaker defines tts_config, it replaces the profile-level config entirely (no merging).
Creating Custom Speakers
- Get Voice IDs from your TTS provider
- Design Personalities that complement each other
- Write Rich Backstories to guide content expertise
- Test Combinations with different content types
๐ Supported Providers
Language Models (via Esperanto)
- OpenAI: GPT-4, GPT-4o, o1, o3
- Anthropic: Claude 3.5 Sonnet, Claude 3 Opus
- Google: Gemini Pro, Gemini Flash
- Groq: Mixtral, Llama models
- Ollama: Local model support
- Perplexity: Research-enhanced models
- Azure OpenAI: Enterprise OpenAI
- Mistral: Mistral models
- DeepSeek: DeepSeek models
- xAI: Grok models
- OpenRouter: Multi-provider access
Text-to-Speech Services
- ElevenLabs: Professional voice synthesis
- OpenAI TTS: High-quality voices
- Google: Google Cloud TTS
- Vertex AI: Google Cloud enterprise
๐ Output Structure
output/episode_name/
โโโ outline.json # Structured outline
โโโ transcript.json # Complete dialogue
โโโ clips/ # Individual audio clips
โ โโโ 0000.mp3 # First segment
โ โโโ 0001.mp3 # Second segment
โ โโโ ... # Additional segments
โโโ audio/ # Final output
โโโ episode_name.mp3 # Complete podcast
๐ ๏ธ CLI Commands
# Launch web interface (requires UI installation)
podcast-creator ui
# Launch on custom port/host
podcast-creator ui --port 8080 --host 0.0.0.0
# Skip dependency check
podcast-creator ui --skip-init-check
# Initialize project with templates
podcast-creator init
# Initialize in specific directory
podcast-creator init --output-dir /path/to/project
# Overwrite existing files
podcast-creator init --force
# Show version
podcast-creator version
Note: The ui command requires the UI installation: pip install podcast-creator[ui]
๐จ Web Interface Features
The podcast-creator ui command launches a comprehensive Streamlit interface that provides:
- ๐ Dashboard: Statistics and quick actions
- ๐๏ธ Speaker Management: Visual profile creation with voice selection dropdowns
- ๐บ Episode Management: Configure generation parameters and AI models
- ๐ฌ Podcast Generation: Multi-content support (text, files, URLs) with real-time progress
- ๐ Episode Library: Audio playback, transcript viewing, and downloads
- ๐ค Import/Export: Share profiles via JSON files
The interface automatically detects missing dependencies and offers to run initialization if needed.
๐ Performance
- โก Parallel Processing: 5 concurrent audio clips per batch (configurable)
- ๐ API-Safe Batching: Respects provider rate limits
- ๐ Scalable: Handles 30+ dialogue segments efficiently
- โฑ๏ธ Fast Generation: ~2-3 minutes for typical podcasts
- ๐ฏ Optimized Workflow: Smart resource management
โ ๏ธ Rate Limiting Configuration
If you encounter errors like ElevenLabs API error: Too many concurrent requests, you can adjust the parallel processing batch size:
# In your .env file
TTS_BATCH_SIZE=2 # Reduce from default 5 to 2 for ElevenLabs free plan
This is particularly useful for:
- ElevenLabs Free Plan: Limited to 2 concurrent requests
- Other TTS providers with stricter rate limits
- Debugging: Set to 1 for sequential processing
๐ Retry Configuration
LLM and TTS API calls automatically retry on transient failures (network errors, timeouts, rate limits) with exponential backoff. Non-retryable errors are raised immediately without retry โ this includes programming errors (e.g. ValueError) and HTTP 4xx client errors (e.g. 404 model not found, 401 auth failure), except 429 rate-limit which is retried.
# In your .env file
PODCAST_RETRY_MAX_ATTEMPTS=3 # Max retry attempts (default: 3)
PODCAST_RETRY_WAIT_MULTIPLIER=5 # Backoff multiplier in seconds (default: 5)
PODCAST_RETRY_WAIT_MAX=30 # Max wait between retries in seconds (default: 30)
You can also configure retries programmatically for LLM calls (outline and transcript generation):
result = await create_podcast(
content="Your content...",
episode_profile="tech_discussion",
episode_name="my_podcast",
output_dir="output/my_podcast",
retry_max_attempts=5, # Override default
retry_wait_multiplier=3, # Override default
)
To disable retries entirely, set PODCAST_RETRY_MAX_ATTEMPTS=1.
๐ Proxy Configuration
If you're behind a corporate firewall or need to route requests through a proxy, use standard environment variables:
# In your .env file or shell environment
HTTP_PROXY=http://proxy.example.com:8080
HTTPS_PROXY=http://proxy.example.com:8080
NO_PROXY=localhost,127.0.0.1
Authenticated Proxies:
# Proxies with authentication are supported
HTTP_PROXY=http://user:password@proxy.example.com:8080
HTTPS_PROXY=http://user:password@proxy.example.com:8080
The underlying libraries (esperanto, content-core) automatically detect and use these standard proxy environment variables for all network requests.
๐งช Development
Installing for Development
git clone <repository-url>
cd podcast-creator
# Install with uv (recommended)
uv sync
# This installs the package in editable mode
# along with all dependencies
Project Structure
podcast-creator/
โโโ src/
โ โโโ podcast_creator/
โ โโโ __init__.py # Public API
โ โโโ config.py # Configuration system
โ โโโ cli.py # CLI commands (with UI command)
โ โโโ core.py # Core utilities
โ โโโ graph.py # LangGraph workflow
โ โโโ nodes.py # Workflow nodes
โ โโโ retry.py # Retry utilities with exponential backoff
โ โโโ speakers.py # Speaker management
โ โโโ episodes.py # Episode profile management
โ โโโ state.py # State management
โ โโโ validators.py # Validation utilities
โ โโโ resources/ # Bundled templates
โ โโโ prompts/
โ โโโ speakers_config.json
โ โโโ episodes_config.json
โ โโโ streamlit_app/ # Web interface
โ โโโ examples/
โโโ pyproject.toml # Package configuration
โโโ README.md
Testing
# Test the package
python -c "from podcast_creator import create_podcast; print('Import successful')"
# Test CLI
podcast-creator --help
# Test web interface
podcast-creator ui
# Test initialization
mkdir test_project
cd test_project
podcast-creator init
python example_usage.py
๐ Examples
Check the examples/ directory for:
- Episode Profiles: Comprehensive guide to streamlined podcast creation
- Basic usage examples
- Advanced configuration
- Custom speaker setups
- Multi-language podcasts
- Different content types
๐ค Contributing
We welcome contributions! Please see our Contributing Guide for details on:
- ๐ Getting started with development
- ๐ Our pull request process
- ๐งช Testing guidelines
- ๐จ Code style and standards
- ๐ How to report bugs
- ๐ก How to suggest new features
Quick links:
๐ License
This project is licensed under the MIT License - see the LICENSE file for details.
๐ Links
- Examples: Examples
Made with โค๏ธ for the AI community
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file podcast_creator-0.12.0.tar.gz.
File metadata
- Download URL: podcast_creator-0.12.0.tar.gz
- Upload date:
- Size: 486.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.7 {"installer":{"name":"uv","version":"0.10.7","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
16efbafb3103fecf7f37f434e8dcaf2e5e2929c505b9be833faa25287bc8d9ad
|
|
| MD5 |
800df5fcba7ec8ee49c8029a52e223de
|
|
| BLAKE2b-256 |
95e3d7112e86ec4194c14f6024dd0a0d1509b5ef70c76bf41c8359db22c6a9d9
|
File details
Details for the file podcast_creator-0.12.0-py3-none-any.whl.
File metadata
- Download URL: podcast_creator-0.12.0-py3-none-any.whl
- Upload date:
- Size: 79.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.7 {"installer":{"name":"uv","version":"0.10.7","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
605080b5841710733b7bcb6ff2fcb05b6a77f2aabd3c7bec617b58b03250fb47
|
|
| MD5 |
c8b2c8d87a5ef9cf3b7ebc31df4d08e3
|
|
| BLAKE2b-256 |
2f6cf9ed9d1e313183506d5774e4490a36ca6c0f231899055d658e283b26aaed
|