A comprehensive tool for validating reference accuracy in academic papers
Project description
RefChecker
Validate reference accuracy in academic papers.
Catch citation errors, fabricated references, and metadata mismatches before they reach reviewers.
Quick Start • Features • Web UI • CLI • Hallucination Detection • Deployment
RefChecker verifies citations against Semantic Scholar, OpenAlex, and CrossRef, and uses LLM-powered web search to flag likely fabricated references. It supports single papers, bulk batches, and automated scanning of entire OpenReview venues.
Built by Mark Russinovich with AI assistants (Cursor, GitHub Copilot, Claude Code). Watch the deep dive video.
Contents
- Quick Start
- Features
- Sample Output
- Install
- Web UI
- CLI
- Hallucination Detection
- Bulk Checking
- OpenReview Integration
- Output & Reports
- Deployment
- Configuration
- Local Database
- Testing
- License
Quick Start
Web UI (Docker)
docker run -p 8000:8000 ghcr.io/markrussinovich/refchecker:latest
Open http://localhost:8000 in your browser.
Web UI (pip)
pip install academic-refchecker[llm,webui]
refchecker-webui
CLI (pip)
pip install academic-refchecker[llm]
academic-refchecker --paper 1706.03762
academic-refchecker --paper /path/to/paper.pdf
Tip: Set
SEMANTIC_SCHOLAR_API_KEYfor 1-2s per reference vs 5-10s without.
Features
| Category | What it does |
|---|---|
| Input formats | ArXiv IDs/URLs, PDFs, LaTeX (.tex), BibTeX (.bib/.bbl), plain text |
| Verification sources | Semantic Scholar, OpenAlex, CrossRef — cross-checked for accuracy |
| LLM extraction | OpenAI, Anthropic, Google, Azure, or local vLLM for parsing complex bibliographies |
| Metadata checks | Titles, authors, years, venues, DOIs, ArXiv IDs, URLs |
| Smart matching | Handles formatting variations (BERT vs B-ERT, pre-trained vs pretrained) |
| Hallucination detection | Flags likely fabricated references using deterministic pre-filters + LLM web search |
| Bulk checking | Upload multiple files or a ZIP in the Web UI; use --paper-list or --openreview in the CLI |
| OpenReview scanning | Fetch all accepted (or submitted) papers for a venue and scan them in one command |
| Reports | JSON, JSONL, CSV, or text — with error details, corrections, and hallucination assessments |
| Corrections | Auto-generates corrected BibTeX, plain-text, and bibitem entries for each error |
| Web UI | Real-time progress, history sidebar, batch tracking, export (Markdown/text/BibTeX), dark mode |
| Multi-user hosting | OAuth sign-in (Google, GitHub, Microsoft), per-user rate limiting, admin controls |
Sample Output
Web UI
CLI — Single Paper
📄 Processing: Attention Is All You Need
URL: https://arxiv.org/abs/1706.03762
[1/45] Neural machine translation in linear time
Nal Kalchbrenner et al. | 2017
⚠️ Warning: Year mismatch: cited '2017', actual '2016'
[2/45] Effective approaches to attention-based neural machine translation
Minh-Thang Luong et al. | 2015
❌ Error: First author mismatch: cited 'Minh-Thang Luong', actual 'Thang Luong'
[3/45] Deep Residual Learning for Image Recognition
Kaiming He et al. | 2016 | https://doi.org/10.1109/CVPR.2016.91
❌ Error: DOI mismatch: cited '10.1109/CVPR.2016.91', actual '10.1109/CVPR.2016.90'
============================================================
📋 SUMMARY
📚 Total references processed: 68
❌ Total errors: 55 ⚠️ Total warnings: 16 ❓ Unverified: 15
CLI — Hallucination Flagging
[5/7] Efficient Neural Network Pruning Using Iterative Sparse Retraining
Shuang Li, Yifan Chen | 2019
❓ Could not verify
🚩 Hallucination assessment: LIKELY
A web search for the exact title and authors yields no results in any
academic database. The paper does not appear in ICML 2019 proceedings,
indicating it is probably fabricated.
Install
PyPI (recommended)
pip install academic-refchecker[llm,webui] # Web UI + CLI + LLM providers
pip install academic-refchecker[llm] # CLI + LLM providers
pip install academic-refchecker # CLI only (regex extraction)
From Source (development)
git clone https://github.com/markrussinovich/refchecker.git && cd refchecker
python -m venv .venv && source .venv/bin/activate # Windows: .venv\Scripts\activate
pip install -e ".[llm,webui]"
pip install -r requirements-dev.txt # pytest, playwright, etc.
Requirements: Python 3.7+ (3.10+ recommended). Node.js 18+ is only needed for Web UI frontend development.
Web UI
The Web UI provides real-time progress, check history, batch tracking, and one-click export of corrections.
refchecker-webui # default: http://localhost:8000
refchecker-webui --port 9000 # custom port
Key features:
- Single check — paste an ArXiv URL/ID or upload a PDF/BibTeX/LaTeX file
- Bulk check — upload multiple files (up to 50) or a single ZIP archive; papers are grouped into a batch with a progress bar
- Bulk URL list — paste up to 50 URLs or ArXiv IDs (one per line) to check in a single batch
- Status dashboard — filterable badge counts for errors, warnings, unverified, and hallucinated references
- Reference cards — per-reference details with corrections, source links (Semantic Scholar, ArXiv, DOI), and hallucination assessment
- Export — download corrections as Markdown, plain text, or BibTeX
- History sidebar — browse and re-run previous checks; batches are grouped together
- Settings — LLM provider/model selection, API key management, dark/light/system theme
Frontend Development
cd web-ui && npm install && npm start # http://localhost:5173
Or run backend and frontend separately:
# Terminal 1 — Backend
python -m uvicorn backend.main:app --reload --port 8000
# Terminal 2 — Frontend
cd web-ui && npm run dev
See web-ui/README.md for more.
CLI
# ArXiv (ID or URL)
academic-refchecker --paper 1706.03762
academic-refchecker --paper https://arxiv.org/abs/1706.03762
# Local files (PDF, LaTeX, text, BibTeX)
academic-refchecker --paper paper.pdf
academic-refchecker --paper paper.tex
academic-refchecker --paper refs.bib
# With LLM extraction (recommended for complex bibliographies)
academic-refchecker --paper paper.pdf --llm-provider anthropic
# Save human-readable output
academic-refchecker --paper 1706.03762 --output-file errors.txt
# Save structured report (JSON, JSONL, CSV, or text)
academic-refchecker --paper 1706.03762 --report-file report.json --report-format json
# Bulk: check a list of papers
academic-refchecker --paper-list papers.txt --report-file report.json
# OpenReview: fetch and scan an entire venue
academic-refchecker --openreview iclr2024 --report-file report.json
All CLI Options
Input (choose one):
--paper PAPER ArXiv ID, URL, PDF, LaTeX, text, or BibTeX file
--paper-list PATH Newline-delimited file of paper specs (URLs, IDs, paths)
--openreview VENUE Fetch papers from an OpenReview venue (e.g. iclr2024)
--openreview-status MODE accepted (default) or submitted
LLM:
--llm-provider PROVIDER openai, anthropic, google, azure, or vllm
--llm-model MODEL Override the default model for the provider
--llm-endpoint URL Custom endpoint (e.g. local vLLM server)
--llm-parallel-chunks Enable parallel LLM chunk processing (default)
--llm-no-parallel-chunks Disable parallel LLM chunk processing
--llm-max-chunk-workers N Max workers for parallel LLM chunks (default: 4)
Verification:
--db-path PATH Path to local Semantic Scholar database
--semantic-scholar-api-key KEY Override SEMANTIC_SCHOLAR_API_KEY env var
--disable-parallel Run verification sequentially
--max-workers N Max parallel verification threads (default: 6)
Output:
--output-file [PATH] Human-readable output (default: reference_errors.txt)
--report-file PATH Structured report (includes hallucination assessments)
--report-format FORMAT json (default), jsonl, csv, or text
--debug Verbose logging
Hallucination Detection
RefChecker automatically evaluates every flagged reference for potential fabrication using a two-stage pipeline.
Stage 1 — Deterministic Pre-filter (no LLM needed)
References are flagged for deeper inspection when they exhibit:
- Unverified status — not found in Semantic Scholar, OpenAlex, or CrossRef
- Author overlap below 60% — fewer than 60% of cited authors match any known paper (applies to references with 3+ authors)
- Identifier conflicts — DOI or ArXiv ID resolves to a different paper
- URL verification failure — cited URL is broken or points to a different paper
References with only minor issues (year off by one, venue variation) are not flagged.
Stage 2 — LLM Web Search Verification
Flagged references are sent to the configured LLM for a web search. The LLM searches for the exact title, authors, and venue to determine whether the paper actually exists.
Each reference receives a verdict:
| Verdict | Meaning |
|---|---|
| 🚩 LIKELY | Probably fabricated — no evidence the paper exists |
| ❓ UNCERTAIN | Inconclusive — may exist but could not be confirmed |
| ✅ UNLIKELY | Real paper — found in academic databases or on the web |
Hallucination assessments appear inline in CLI output, in Web UI reference cards, and in structured reports (JSON/JSONL/CSV) via the hallucination_assessment field.
Bulk Checking
Web UI
Upload multiple files or a ZIP archive to check up to 50 papers in a single batch. Alternatively, paste a list of URLs or ArXiv IDs (one per line). Batches track progress per paper and appear as a group in the history sidebar.
Supported file types: PDF, TXT, TEX, BIB, BBL, ZIP.
CLI
Create a text file with one paper per line (ArXiv IDs, URLs, or local file paths):
1706.03762
https://openreview.net/pdf?id=ZG3RaNIsO8
paper/local_sample.bib
/path/to/paper.pdf
Then run:
academic-refchecker --paper-list papers.txt --report-file bulk_report.json
The report includes per-paper rollups and a cross-paper summary with flagged reference counts.
OpenReview Integration
Scan all accepted (or submitted) papers for an OpenReview venue in one command:
# Scan accepted papers
academic-refchecker --openreview iclr2024 --report-file report.json
# Scan all public submissions instead
academic-refchecker --openreview iclr2024 --openreview-status submitted --report-file report.json
Supported venues include: ICLR, NeurIPS, ICML, AISTATS, AAAI, IJCAI — use formats like iclr2024, NeurIPS-2023, or neurips_2024.
The command fetches the paper list from the OpenReview API, writes it to output/openreview_<venue>_<status>.txt, and then runs a bulk scan. The structured report includes per-paper rollups with flagged record counts and error-type distributions, making it easy to triage an entire conference for citation problems.
Output & Reports
Result Types
| Type | Description | Examples |
|---|---|---|
| ❌ Error | Critical issues needing correction | Author/title/DOI mismatches, incorrect ArXiv IDs |
| ⚠️ Warning | Minor issues to review | Year differences, venue variations |
| ℹ️ Suggestion | Recommended improvements | Add missing ArXiv/DOI URLs |
| ❓ Unverified | Could not verify against any source | Rare publications, preprints |
| 🚩 Hallucination | Likely fabricated reference | Unverifiable with rich metadata, identifier conflicts |
Structured Reports
Write machine-readable reports with --report-file and --report-format:
academic-refchecker --paper 1706.03762 --report-file report.json --report-format json
Example JSON report structure
{
"generated_at": "2026-03-15T19:50:52Z",
"summary": {
"total_papers_processed": 1,
"total_references_processed": 7,
"total_errors_found": 2,
"total_warnings_found": 2,
"total_unverified_refs": 4,
"flagged_records": 3,
"flagged_papers": 1
},
"papers": [
{
"source_paper_id": "local_hallucination_7ref_sample",
"source_title": "Hallucination 7Ref Sample",
"total_records": 6,
"flagged_records": 3,
"max_flag_level": "high",
"error_type_counts": { "unverified": 3, "multiple": 2, "year (v1 vs v2 update)": 1 },
"reason_counts": { "unverified": 3, "web_search_not_found": 3 }
}
],
"records": [
{
"ref_title": "Deep Residual Learning for Image Recognition",
"ref_authors_cited": "Jian He, Xiangyu Zhang, Shaoqing Ren, Jian Sun",
"ref_authors_correct": "Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun",
"error_type": "multiple",
"error_details": "- First author mismatch ...\n- Year mismatch ...",
"ref_corrected_bibtex": "@inproceedings{he2016resnet, ... year = {2015} ...}",
"hallucination_assessment": { "verdict": "UNLIKELY", "explanation": "..." }
}
]
}
CLI output examples
❌ Error: First author mismatch: cited 'Jian He', actual 'Kaiming He'
❌ Error: DOI mismatch: cited '10.5555/3295222.3295349', actual '10.48550/arXiv.1706.03762'
⚠️ Warning: Year mismatch: cited '2019', actual '2018'
ℹ️ Suggestion: Add ArXiv URL https://arxiv.org/abs/1706.03762
❓ Could not verify: Llama guard (M. A. Research, 2024)
🚩 Hallucination assessment: LIKELY — no matching paper found in academic databases
Each report record includes the original reference, error details, corrected metadata (BibTeX, plain text, bibitem), verified URLs, and hallucination assessment when applicable.
Deployment
Docker
Pre-built multi-architecture images are published to GitHub Container Registry on every release.
# Quick start
docker run -p 8000:8000 ghcr.io/markrussinovich/refchecker:latest
# With LLM API key (recommended)
docker run -p 8000:8000 -e ANTHROPIC_API_KEY=your_key ghcr.io/markrussinovich/refchecker:latest
# Persistent data
docker run -p 8000:8000 \
-e ANTHROPIC_API_KEY=your_key \
-v refchecker-data:/app/data \
ghcr.io/markrussinovich/refchecker:latest
Other LLM providers:
docker run -p 8000:8000 -e OPENAI_API_KEY=your_key ghcr.io/markrussinovich/refchecker:latest
docker run -p 8000:8000 -e GOOGLE_API_KEY=your_key ghcr.io/markrussinovich/refchecker:latest
Docker Compose
git clone https://github.com/markrussinovich/refchecker.git && cd refchecker
cp .env.example .env # Add your API keys
docker compose up -d
docker compose logs -f # View logs
docker compose down # Stop
docker compose pull # Update to latest
| Tag | Description | Arch | Size |
|---|---|---|---|
latest |
Latest stable release | amd64, arm64 | ~800MB |
X.Y.Z |
Specific version (e.g., 2.0.18) |
amd64, arm64 | ~800MB |
Multi-User Server (OAuth)
By default, RefChecker runs in single-user mode — no login required. Enable multi-user mode for shared deployments where each visitor signs in via OAuth. LLM API keys are entered per-user in the Settings panel, stored in the browser's localStorage, and sent per-request — never stored on the server.
1. Generate a JWT Secret Key
python -c "import secrets; print(secrets.token_hex(32))"
2. Register an OAuth Application
Configure at least one provider:
| Provider | Registration URL | Callback URL |
|---|---|---|
| Google Cloud Console | https://<domain>/api/auth/callback/google |
|
| GitHub | GitHub Developer Settings | https://<domain>/api/auth/callback/github |
| Microsoft | Azure App Registrations | https://<domain>/api/auth/callback/microsoft |
3. Configure Environment Variables
cp .env.example .env
REFCHECKER_MULTIUSER=true
JWT_SECRET_KEY=<output from step 1>
SITE_URL=https://<your-domain>
HTTPS_ONLY=true
# At least one OAuth provider
GOOGLE_CLIENT_ID=...
GOOGLE_CLIENT_SECRET=...
GITHUB_CLIENT_ID=...
GITHUB_CLIENT_SECRET=...
MS_CLIENT_ID=...
MS_CLIENT_SECRET=...
# Optional
REFCHECKER_ADMINS=github:you # comma-separated; first sign-in is auto-admin
MAX_CHECKS_PER_USER=3 # max concurrent checks per user (default: 3)
4. Launch
docker compose up -d
Or without Docker:
pip install "academic-refchecker[llm,webui]"
REFCHECKER_MULTIUSER=true JWT_SECRET_KEY=<secret> GOOGLE_CLIENT_ID=... GOOGLE_CLIENT_SECRET=... \
refchecker-webui --port 8000
Verify:
curl http://localhost:8000/api/auth/providers
# {"providers":["google","github"]}
Notes:
- The first user to sign in is automatically admin. Add more via
REFCHECKER_ADMINS. - Each user may run up to
MAX_CHECKS_PER_USERconcurrent checks (default 3). The 4th returns HTTP 429. - The CLI is unaffected —
academic-refcheckerworks without any auth configuration. - Place the server behind a TLS-terminating reverse proxy (nginx, Caddy) for HTTPS.
Deploy to Render
RefChecker includes a render.yaml Blueprint for one-click deployment to Render:
- Fork this repo (or connect your own copy).
- On Render, click New + → Blueprint → select the repo.
- Render reads
render.yamland creates the service with a persistent disk. - Set environment variables in the Render dashboard (Environment tab):
SITE_URL— your public URL includinghttps://(must match exactly — OAuth fails otherwise).HTTPS_ONLY=truefor production.REFCHECKER_DATA_DIR=/data(matches the persistent disk mount).- At least one OAuth provider's
CLIENT_ID/CLIENT_SECRET.
- Register each provider's callback URL as
https://<your-url>/api/auth/callback/{google,github,microsoft}.
Note: The persistent disk at
/datastores the SQLite database and uploaded files, so data survives redeployments. For other PaaS hosts (Railway, Fly.io), the same Docker image works — setPORT,REFCHECKER_DATA_DIR, and the auth env vars.
Configuration
LLM Providers
LLM-powered extraction improves accuracy with complex bibliographies. Claude Sonnet 4 performs best; GPT-4o may hallucinate DOIs.
| Provider | Env Variable | Example Model |
|---|---|---|
| Anthropic | ANTHROPIC_API_KEY |
claude-sonnet-4-20250514 |
| OpenAI | OPENAI_API_KEY |
gpt-5.2-mini |
GOOGLE_API_KEY |
gemini-3 |
|
| Azure | AZURE_OPENAI_API_KEY |
gpt-4o |
| vLLM | (local) | meta-llama/Llama-3.3-70B-Instruct |
export ANTHROPIC_API_KEY=your_key
academic-refchecker --paper 1706.03762 --llm-provider anthropic
academic-refchecker --paper paper.pdf --llm-provider openai --llm-model gpt-5.2-mini
academic-refchecker --paper paper.pdf --llm-provider vllm --llm-model meta-llama/Llama-3.3-70B-Instruct
Local Models (vLLM)
Run an OpenAI-compatible vLLM server for local inference:
pip install "academic-refchecker[vllm]"
python scripts/start_vllm_server.py --model meta-llama/Llama-3.3-70B-Instruct --port 8001
academic-refchecker --paper paper.pdf --llm-provider vllm --llm-endpoint http://localhost:8001/v1
Environment Variables
# LLM
export REFCHECKER_LLM_PROVIDER=anthropic
export ANTHROPIC_API_KEY=your_key # Also: OPENAI_API_KEY, GOOGLE_API_KEY
# Performance
export SEMANTIC_SCHOLAR_API_KEY=your_key # Higher rate limits / faster verification
Local Database
For offline verification or faster processing:
python scripts/download_db.py \
--field "computer science" \
--start-year 2020 --end-year 2024
academic-refchecker --paper paper.pdf --db-path semantic_scholar_db/semantic_scholar.db
Testing
680+ tests covering unit, integration, and end-to-end scenarios.
pytest tests/ # All tests
pytest tests/unit/ # Unit only
pytest tests/e2e/ # End-to-end (Playwright)
pytest --cov=src tests/ # With coverage
See tests/README.md for details.
License
MIT License — see LICENSE.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file academic_refchecker-3.0.79.tar.gz.
File metadata
- Download URL: academic_refchecker-3.0.79.tar.gz
- Upload date:
- Size: 659.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.15
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
bab508733c81ab31f24448cb6a8013e403742c9aad1cb9c968bef975f98c7ffa
|
|
| MD5 |
476d5f82fb8159efc0a0846a482b368f
|
|
| BLAKE2b-256 |
e1d3a5e308aa92e7a032d2a72482e3662a81d63486e70bf3ebf963d5e33efa9d
|
File details
Details for the file academic_refchecker-3.0.79-py3-none-any.whl.
File metadata
- Download URL: academic_refchecker-3.0.79-py3-none-any.whl
- Upload date:
- Size: 683.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.15
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
91653616fa85e735d4d631a2bce0729d16d645abdb7a9e11e031f37d6695feee
|
|
| MD5 |
70a21b54e5ecdb4c5f5aafbf536e05a1
|
|
| BLAKE2b-256 |
5f4ca365c364c3c0201eb9cacff44ef95a2fa63546c05d0e7ad457a87264bd60
|