Skip to main content

AI-powered academic paper reviewer

Project description

OpenAIReview

PyPI version

Our goal is provide thorough and detailed reviews to help researchers conduct the best research. See more examples here.

Example

Installation

uv venv && uv pip install openaireview
# or: pip install openaireview

For development:

git clone https://github.com/ChicagoHAI/OpenAIReview.git
cd OpenAIReview
uv venv && uv pip install -e .
# or: pip install -e .

PDF math support (optional)

For math-heavy PDFs, install Marker separately to get accurate LaTeX extraction. Without Marker, PDFs are processed with PyMuPDF which cannot extract math symbols correctly.

# Install Marker CLI in an isolated environment (avoids dependency conflicts)
uv tool install marker-pdf --with psutil

Marker is used automatically when available on PATH. It is most useful for math-heavy PDFs, but runs very slowly without a GPU. For papers with math, we recommend using .tex source, .md, or arXiv HTML URLs instead of PDF when possible — these always produce correct output without needing Marker.

Quick Start

First, set your OpenRouter API key (get one at openrouter.ai/keys):

export OPENROUTER_API_KEY=your_key_here

Or create a .env file in your working directory:

OPENROUTER_API_KEY=your_key_here

Then review a paper and visualize results:

# Review a local file
openaireview review paper.pdf

# Or review directly from an arXiv URL
openaireview review https://arxiv.org/html/2602.18458v1

# Visualize results
openaireview serve
# Open http://localhost:8080

CLI Reference

openaireview review <file_or_url>

Review an academic paper for technical and logical issues. Accepts a local file path or an arXiv URL.

Option Default Description
--method incremental Review method: zero_shot, local, incremental, incremental_full
--model anthropic/claude-opus-4-6 Model to use
--output-dir ./review_results Directory for output JSON files
--name (from filename) Paper slug name

openaireview serve

Start a local visualization server to browse review results.

Option Default Description
--results-dir ./review_results Directory containing result JSON files
--port 8080 Server port

Supported Input Formats

  • PDF (.pdf) — uses Marker for high-quality extraction with LaTeX math; falls back to PyMuPDF if Marker is not installed
  • DOCX (.docx) — via python-docx
  • LaTeX (.tex) — plain text with title extraction from \title{}
  • Text/Markdown (.txt, .md) — plain text
  • arXiv HTML — fetch and parse directly from https://arxiv.org/html/<id> or https://arxiv.org/abs/<id>

Environment Variables

Variable Default Description
OPENROUTER_API_KEY (required) Your OpenRouter API key
MODEL anthropic/claude-opus-4-6 Default model

These can be set as environment variables or in a .env file. See .env.example for a template.

Supported Models & Pricing

All models available on OpenRouter are supported — use any model ID via --model. The following models have built-in pricing for accurate cost tracking in the visualization:

Model Input ($/1M tokens) Output ($/1M tokens)
anthropic/claude-opus-4-6 $5.00 $25.00
anthropic/claude-opus-4-5 $5.00 $25.00
openai/gpt-5.2-pro $21.00 $168.00
google/gemini-3.1-pro-preview $2.00 $12.00

For models not listed above, a default rate of $5.00/$25.00 per 1M tokens is used.

Review Methods

  • zero_shot — single prompt asking the model to find all issues
  • local — deep-checks each chunk with surrounding window context (no filtering)
  • incremental — sequential processing with running summary, then consolidation
  • incremental_full — same as incremental but returns all comments before consolidation

Benchmarks

Benchmark data and experiment scripts are in benchmarks/. See benchmarks/REPORT.md for results.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

openaireview-0.1.5.tar.gz (30.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

openaireview-0.1.5-py3-none-any.whl (36.5 kB view details)

Uploaded Python 3

File details

Details for the file openaireview-0.1.5.tar.gz.

File metadata

  • Download URL: openaireview-0.1.5.tar.gz
  • Upload date:
  • Size: 30.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for openaireview-0.1.5.tar.gz
Algorithm Hash digest
SHA256 c4617987a7e681a27a7dab87819836a17afdc03b3d3fd003e27d4196a1e27975
MD5 2e3d2c3215ad9829e9c40ba68839b088
BLAKE2b-256 8e6c0da9ce6ca4a025d8d0ee2dba8e9de31ecfc0504510415c2e9576066691d9

See more details on using hashes here.

Provenance

The following attestation bundles were made for openaireview-0.1.5.tar.gz:

Publisher: publish.yml on ChicagoHAI/OpenAIReview

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file openaireview-0.1.5-py3-none-any.whl.

File metadata

  • Download URL: openaireview-0.1.5-py3-none-any.whl
  • Upload date:
  • Size: 36.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for openaireview-0.1.5-py3-none-any.whl
Algorithm Hash digest
SHA256 e20eed91010aaf2bd3126244f6438bde17c9bb7473bb1387e333269ac13004b9
MD5 374e6e925a74d24a2bfe41ac11debe82
BLAKE2b-256 6410178f3b790d81b9def8bc1c6109963bd2abbca781a01a75ce35032936e3ea

See more details on using hashes here.

Provenance

The following attestation bundles were made for openaireview-0.1.5-py3-none-any.whl:

Publisher: publish.yml on ChicagoHAI/OpenAIReview

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page