Multi-source speech transcription with LLM-based adjudication inspired by textual criticism
Reason this release was yanked:
Name clash on GitHub
Project description
Transcript Critic
Automated pipeline for producing accurate speech transcripts from video URLs. Downloads media, transcribes with multiple Whisper models, and merges all available sources — Whisper, YouTube captions, and optional external transcripts — into a single "critical text" using LLM-based adjudication.
The approach applies principles from textual criticism: multiple independent "witnesses" to the same speech are aligned, compared, and merged by an LLM that judges each difference on its merits, without knowing which source produced which reading. This builds on earlier work applying similar techniques to OCR (Ringger & Lund, 2014; Lund et al., 2013), replacing trained classifiers with an LLM as the eclectic editor.
How is this different from WhisperX?
WhisperX improves a single Whisper run with voice-activity-detection (VAD) chunking, word-level timestamps, and speaker diarization — but the transcript still comes from one model pass. Transcript Critic takes a different approach: it runs multiple Whisper models, pulls in YouTube captions and external human-edited transcripts, and treats them all as independent witnesses. An LLM then adjudicates every disagreement blindly, without knowing which source produced which reading. The result is a merged "critical text" that is more accurate than any single source. If you just need fast, well-segmented Whisper output, WhisperX is the right tool; if you want the most accurate transcript possible from multiple sources, this is.
Features
- Critical text merging: Combines 2–3+ transcript sources into the most accurate version using blind, anonymous presentation to an LLM — no source receives preferential treatment
- wdiff-based alignment: Uses longest common subsequence alignment (via
wdiff) to keep chunks properly aligned across sources of different lengths, replacing naive proportional slicing - Multi-model Whisper ensembling: Runs multiple Whisper models (e.g., small + medium) and resolves disagreements via LLM
- External transcript support: Merges in human-edited transcripts (e.g., from publisher websites) as an additional source
- Structured transcript preservation: When external transcripts have speaker labels and timestamps, the merged output preserves that structure
- Slide extraction and analysis: Automatic scene detection for presentation slides, with optional vision API descriptions
- Make-style DAG pipeline: Each stage checks whether its outputs are newer than its inputs, skipping unnecessary work
- Checkpoint resumption: Long operations save checkpoints and resume after interruption — merge chunks, diarization segmentation, and embedding extraction all checkpoint independently
- Cost estimation: Shows estimated API costs before running (
--dry-runfor estimation only) - Local-first LLM: Uses Ollama by default for free, local operation — no API key needed
- Speaker diarization: Identifies who is speaking using pyannote.audio, with automatic or manual speaker naming — LLM speaker identification uses video metadata (title, description) for correct name spellings
- Timestamped logging: All pipeline output prefixed with
[HH:MM:SS]wall-clock timestamps for log correlation during long runs - Whisper-only mode:
--no-llmto skip all LLM features and run Whisper only
Installation
pip install transcript-critic
System Dependencies
# Required tools
brew install ffmpeg wdiff # macOS
# apt install ffmpeg wdiff # Ubuntu/Debian
# Install Ollama for local LLM (used by default for merging/ensembling)
brew install ollama # macOS
# curl -fsSL https://ollama.com/install.sh | sh # Linux
# Pull a model (one-time)
ollama pull qwen2.5
From Source
git clone https://github.com/ringger/transcript-critic.git
cd transcript-critic
pip install -e . # editable install
pip install -e .[dev] # with test dependencies
pip install -e .[diarize] # with speaker diarization
Quick Start
# Basic: Whisper transcription + local LLM merge (free, uses Ollama)
transcript-critic "https://youtube.com/watch?v=..."
# With an external human-edited transcript for three-way merge
transcript-critic "https://youtube.com/watch?v=..." \
--external-transcript "https://example.com/transcript"
# Use Anthropic Claude API instead of local Ollama (higher quality, costs money)
transcript-critic "https://youtube.com/watch?v=..." --api
# Whisper only — no LLM merging at all
transcript-critic "https://youtube.com/watch?v=..." --no-llm
Usage Examples
Podcast
# Podcast episode — audio only, no video or captions
transcript-critic --podcast "https://www.iheart.com/podcast/.../episode/..."
transcript-critic --podcast "https://podcasts.apple.com/us/podcast/..."
Speaker Diarization
# Identify who is speaking (requires pyannote.audio and HF_TOKEN)
pip install pyannote.audio
export HF_TOKEN="hf_..." # HuggingFace token with pyannote model access
# Auto-detect speaker names from introductions
transcript-critic --diarize --num-speakers 2 --podcast "https://..."
# Manual speaker names (in order of first appearance)
transcript-critic --diarize --speaker-names "Ross Douthat,Dario Amodei" --podcast "https://..."
Speech-Only (No Slides)
# YouTube talk or interview — skip slide extraction
transcript-critic "https://youtube.com/watch?v=..." --no-slides
# With external transcript for higher accuracy
transcript-critic "https://youtube.com/watch?v=..." \
--no-slides \
--external-transcript "https://example.com/transcript"
Presentation with Slides
# Extract slides and interleave with transcript
transcript-critic "https://youtube.com/watch?v=..."
# Also describe slide content with vision API
transcript-critic "https://youtube.com/watch?v=..." --analyze-slides
Custom Options
# Custom output directory
transcript-critic "https://youtube.com/watch?v=..." -o ./my_transcript
# Use specific Whisper models
transcript-critic "https://youtube.com/watch?v=..." --whisper-models large
# Use a different local model
transcript-critic "https://youtube.com/watch?v=..." --local-model llama3.3
# Adjust slide detection sensitivity (0.0–1.0, lower = more slides)
transcript-critic "https://youtube.com/watch?v=..." --scene-threshold 0.15
# Force re-processing (ignore existing files)
transcript-critic "https://youtube.com/watch?v=..." --force
# Verbose output
transcript-critic "https://youtube.com/watch?v=..." -v
Output Files
output_dir/
├── metadata.json # Source URL, title, duration, etc.
├── audio.mp3 # Downloaded audio
├── video.mp4 # Downloaded video (if slides enabled)
├── captions.en.vtt # YouTube captions (if available)
├── small.txt # Whisper small transcript
├── medium.txt # Whisper medium transcript
├── ensembled.txt # Ensembled from multiple Whisper models
├── medium.json # Transcript with timestamps
├── diarization.json # Speaker segments (if --diarize)
├── diarization_segmentation.npy # Cached segmentation (if --diarize)
├── diarization_embeddings.npy # Cached embeddings (if --diarize)
├── diarized.txt # Speaker-labeled transcript (if --diarize)
├── transcript_merged.txt # Critical text (merged from all sources)
├── analysis.md # Source survival analysis
├── transcript.md # Final markdown output
├── merge_chunks/ # Per-chunk checkpoints (resumable)
│ ├── .version
│ ├── chunk_000.json
│ └── ...
├── slide_timestamps.json # Slide timing data
├── slides_transcript.json # (if --analyze-slides)
└── slides/ # (if slides enabled)
├── slide_0001.png
└── ...
Pipeline Stages
Optional stages are skipped based on flags. Stage numbers are fixed regardless of which stages run.
| Stage | Tool | Optional |
|---|---|---|
| [1] Download media | yt-dlp | No |
| [2] Transcribe audio | mlx-whisper | No |
| [2b] Speaker diarization | pyannote.audio | Yes (--diarize) |
| [3] Extract slides | ffmpeg | Yes (skipped with --no-slides / --podcast) |
| [4] Analyze slides with vision | LLM + vision | Yes (--analyze-slides) |
| [4b] Merge transcript sources | LLM + wdiff | Yes (on by default; --no-merge to skip) |
| [5] Generate markdown | Python | No |
| [6] Source survival analysis | wdiff | No |
How It Works
Critical Text Merging
The core idea — inspired by textual criticism — is to treat multiple transcripts as independent witnesses to the same speech and adjudicate their differences. Given 2–3+ sources:
- Align all sources against an anchor text using
wdiff(longest common subsequence), producing word-position maps that keep chunks synchronized even when sources differ in length - Chunk the aligned sources into ~500-word segments
- Present each chunk to Claude with anonymous labels (Source 1, Source 2, Source 3) — source names are never revealed, preventing provenance bias
- Adjudicate — Claude chooses the best reading at each point of disagreement, preferring proper nouns, grammatical correctness, and contextual fit
- Reassemble the merged chunks, restoring speaker labels and timestamps from the structured source (if available)
When an external transcript has structure (speaker labels, timestamps), the merge preserves that skeleton while improving the text content from all sources.
Unlike a traditional critical edition, the pipeline does not produce an apparatus of variants, construct a stemma of source relationships, or preserve editorial rationale for each decision. The goal is a single best-reading transcript, not a scholarly edition.
Source Survival Analysis
After merging, wdiff -s compares each source against the merged output, showing how much each source contributed to the final text. Here is an actual survival analysis from a 3-hour podcast episode transcribed with Whisper (small + medium ensembled), YouTube auto-captions, and a human-edited external transcript:
Source Words Common % of Merged % of Source
------------------------- -------- -------- ------------ ------------
Whisper (ensembled) 28,277 27,441 90% 97%
YouTube captions 30,668 28,741 94% 94%
External transcript 33,122 30,245 99% 91%
Merged output 30,524
No single source matches the merged output — the merged text draws from all three. The external transcript is closest (99% of merged words present), but the merge still corrects ~1% of its content using the other sources. Whisper contributes readings not found in either captions or the external transcript, and vice versa.
Here are specific corrections the merge made by adjudicating across sources:
| Whisper | YouTube captions | External transcript | Merged (correct) |
|---|---|---|---|
| "Cloud Opus" | — | "Claude Opus" | Claude Opus (product name) |
| "Ross Douthend" | "ross douthat" | "Ross Douthat" | Ross Douthat (person name) |
| "GPT 5.3 codecs" | — | "GPT-5.3 Codex" | GPT 5.3 Codex (model name, not audio codec) |
| "is source code" | — | "its source code" | its source code (grammar) |
Each source alone gets some things right and others wrong. Whisper hallucinates proper nouns ("Cloud" for "Claude", "Douthend" for "Douthat"). YouTube captions lack capitalization and punctuation but sometimes have correct spellings. The external transcript has the best proper nouns but may paraphrase or omit filler words. The merge selects the best reading at each disagreement, producing a transcript more accurate than any individual source.
Multi-Model Ensembling
When using multiple Whisper models (default: small,medium):
- Runs each model independently
- Uses
wdiffto identify differences (normalized: no caps, no punctuation) - Claude resolves disagreements, preferring real words over transcription errors and proper nouns over generic alternatives
Ensembling is part of the same witness-and-adjudicate process as the critical text merge — multiple Whisper models are simply additional witnesses alongside captions and external transcripts. The implementation runs Whisper-vs-Whisper adjudication first to produce a single ensembled witness, but the principle is the same throughout.
Speaker Diarization
When --diarize is enabled, the pipeline identifies who is speaking at each point in the audio by combining two independent signals:
-
pyannote.audio runs a neural segmentation model over the audio in sliding ~5-second windows, producing frame-level speaker activity probabilities. A global clustering step stitches local predictions across the full recording into consistent speaker labels (SPEAKER_00, SPEAKER_01, etc.). The model handles overlapping speech natively and operates purely on the audio signal — no linguistic content is used.
-
Whisper word timestamps (
--word-timestamps True) provide per-word{start, end}timing from the transcription model.
The pipeline links these by midpoint matching: for each word, it finds which speaker segment overlaps the word's temporal midpoint. Each transcript segment is then assigned the majority speaker of its constituent words. The result is a structured transcript in bracketed format ([H:MM:SS] Speaker: text) that feeds directly into the existing merge pipeline as a structural skeleton.
Speaker identification maps generic labels to real names via three methods (in priority order):
--speaker-names "Alice,Bob"— manual mapping by order of first appearance- LLM-based detection — reads the first ~500 words and infers names from introductions, using video metadata (title, description, channel) for correct spellings (e.g., corrects Whisper's "Douthit" to "Douthat" when the video description contains the correct name)
--no-llm— keeps generic SPEAKER_00/SPEAKER_01 labels
Diarization checkpointing breaks the expensive pyannote pipeline into 6 independently cached steps. Segmentation (the neural model pass, ~50% of runtime) and embedding extraction (the other slow step) both save to .npy files. Embeddings checkpoint every 10 batches to a partial file, enabling resume mid-extraction. If any step's output is newer than the audio file, it is skipped on re-run.
Make-Style Staleness Checks
Every stage checks is_up_to_date(output, *inputs) — if the output file is newer than all input files, the stage is skipped. This means you can re-run the pipeline after changing options and only the affected stages will execute.
Cost Estimation
==================================================
ESTIMATED API COSTS
==================================================
Source merging: 3 sources × 59 chunks = $1.03
Whisper ensemble: 2 models × 59 chunks = $0.92
TOTAL: $1.95 (estimate)
==================================================
Typical Costs
| Feature | 20-min speech | 3-hour podcast |
|---|---|---|
| Whisper ensemble | $0.05–$0.15 | $0.50–$1.00 |
| Source merging (2 sources) | $0.10–$0.30 | $0.50–$1.00 |
| Source merging (3 sources) | $0.15–$0.40 | $1.00–$2.00 |
| Slide analysis | $0.50–$2.00 | N/A |
| Local Ollama (default) | Free | Free |
--no-llm |
Free | Free |
Background
This tool is inspired by textual criticism — the scholarly discipline of comparing multiple manuscript witnesses to reconstruct an authoritative text — applying its core principles (independent witnesses, alignment, adjudication) to speech transcription.
The approach has roots in earlier work applying noisy-channel models and multi-source correction to speech and OCR:
- Ringger & Allen (1996) — Error Correction via a Post-Processor for Continuous Speech Recognition (ICASSP). Introduced SpeechPP, a noisy-channel post-processor that corrects ASR output using language and channel models with Viterbi beam search, developed as part of the TRAINS/TRIPS spoken dialogue systems at the University of Rochester. Extended with a fertility channel model in Ringger & Allen, ICSLP 1996.
- Ringger & Lund (2014) — How Well Does Multiple OCR Error Correction Generalize? Demonstrated that aligning and merging outputs from multiple OCR engines significantly reduces word error rates.
- Lund et al. (2013) — Error Correction with In-Domain Training Across Multiple OCR System Outputs. Used A* alignment and trained classifiers (CRFs, MaxEnt) to choose the best reading from multiple OCR witnesses — a 52% relative decrease in word error rate.
The OCR work used A* alignment because page layout provides natural line boundaries, making alignment a series of short, bounded search problems. Speech has no such boundaries — different ASR systems segment a continuous audio stream arbitrarily — so this tool uses wdiff (LCS-based global alignment) instead. It also replaces the trained classifiers with an LLM, which brings world knowledge and contextual reasoning without requiring task-specific training data. The blind/anonymous presentation of sources is borrowed from peer review and prevents the LLM from developing source-level biases.
Related work in speech:
- ROVER (Fiscus, 1997) — Statistical voting across multiple ASR outputs via word transition networks
- Ensemble Methods for ASR (Lehmann) — Random Forest classifier for selecting words from multiple ASR systems
Troubleshooting
"No Whisper implementation found"
pip install mlx-whisper # Apple Silicon (recommended)
pip install openai-whisper # Other platforms
wdiff not found
Required for alignment-based merging:
brew install wdiff # macOS
apt install wdiff # Ubuntu/Debian
Diarization fails on short audio clips
pyannote's audio decoder can produce sample-count mismatches with MP3 files, especially short clips. Place a WAV version of the audio alongside the MP3:
ffmpeg -i output_dir/audio.mp3 -ar 16000 -ac 1 output_dir/audio.wav
The pipeline automatically prefers audio.wav over audio.mp3 for diarization when both exist.
API timeouts
The tool retries on timeouts (120s per attempt, up to 5 retries with exponential backoff). Long merges save per-chunk checkpoints, so interrupted runs resume from the last completed chunk.
ffmpeg scene detection captures too few/many slides
transcript-critic "..." --scene-threshold 0.05 # More slides
transcript-critic "..." --scene-threshold 0.20 # Fewer slides
License
MIT
Acknowledgments
- OpenAI Whisper — Speech recognition
- MLX Whisper — Apple Silicon optimization
- yt-dlp — Media downloading
- Anthropic Claude — LLM-based adjudication and vision analysis
- pyannote.audio — Speaker diarization
- wdiff — Word-level diff for alignment and comparison
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file transcript_critic-0.1.0.tar.gz.
File metadata
- Download URL: transcript_critic-0.1.0.tar.gz
- Upload date:
- Size: 86.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7d7f13ae158dcb0b70de706bacf96671d8c09211888c4260dc648054d02c36c3
|
|
| MD5 |
2d138b51b76008dc8b4dff1c127cfa9b
|
|
| BLAKE2b-256 |
0212239a11393f5e10bb2b60b83007f591e6657d574809f9dafe9179eab50e24
|
File details
Details for the file transcript_critic-0.1.0-py3-none-any.whl.
File metadata
- Download URL: transcript_critic-0.1.0-py3-none-any.whl
- Upload date:
- Size: 51.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7383e0455e0370ed5ac0f5ed217c68a0c080131f37849ad1ef170022d783365b
|
|
| MD5 |
f2e93e7f1b14aa1c97fe6a7f1c87db3f
|
|
| BLAKE2b-256 |
6865ba436c41acf3a23f18ef691d6e30269596cdb9c703d7b79c418fc7084ed1
|