Unified paper database for coding agents + PaperQA2
Project description
paperpipe
The problem: You're implementing a paper. You need the exact equations, want to verify your code matches the math, and your coding agent keeps hallucinating details. Reading PDFs is slow; copy-pasting LaTeX is tedious.
The solution: paperpipe maintains a local paper database with PDFs, LaTeX source (when available), extracted equations, and coding-oriented summaries. It integrates with coding agents (Claude Code, Codex, Gemini CLI) so they can ground their responses in actual paper content.
Typical workflow
# 1. Add papers you're implementing
papi add 2303.08813 # LoRA paper
papi add https://arxiv.org/abs/1706.03762 # Attention paper
# 2. Check what equations you need to implement
papi show lora --level eq # prints equations to stdout
# 3. Verify your code matches the paper
# (or let your coding agent do this via the /papi skill)
papi show lora --level tex # exact LaTeX definitions
# 4. Ask cross-paper questions (requires RAG backend)
papi ask "How does LoRA differ from full fine-tuning in terms of parameter count?"
# 5. Keep implementation notes
papi notes lora # opens notes.md in $EDITOR
Installation
# Basic (uv recommended)
uv tool install paperpipe
# With features
uv tool install paperpipe --with "paperpipe[llm]" # better summaries via LLMs
uv tool install paperpipe --with "paperpipe[paperqa]" # RAG via PaperQA2
uv tool install paperpipe --with "paperpipe[leann]" # local RAG via LEANN
uv tool install paperpipe --with "paperpipe[figures]" # figure extraction from LaTeX/PDF
uv tool install paperpipe --with "paperpipe[mcp]" # MCP server integrations (Python 3.11+)
uv tool install paperpipe --with "paperpipe[all]" # everything
Alternative: pip install
pip install paperpipe
pip install 'paperpipe[llm]'
pip install 'paperpipe[paperqa]' # PaperQA2 + multimodal PDF parsing
pip install 'paperpipe[leann]'
pip install 'paperpipe[figures]' # figure extraction from LaTeX/PDF
pip install 'paperpipe[mcp]'
pip install 'paperpipe[all]'
From source
git clone https://github.com/hummat/paperpipe && cd paperpipe
pip install -e ".[all]"
What paperpipe stores
~/.paperpipe/ # override with PAPER_DB_PATH
├── index.json
├── .pqa_papers/ # staged PDFs for RAG (created on first `papi ask`)
├── .pqa_index/ # PaperQA2 index cache
├── .leann/ # LEANN index cache
├── papers/
│ └── lora/
│ ├── paper.pdf # for RAG backends
│ ├── source.tex # full LaTeX (if available from arXiv)
│ ├── equations.md # extracted equations with context
│ ├── summary.md # coding-oriented summary
│ ├── tldr.md # one-paragraph TL;DR
│ ├── meta.json # metadata + tags
│ ├── notes.md # your implementation notes
│ └── figures/ # extracted figures (if available)
│ ├── figure1.png
│ └── figure2.pdf
Why this structure matters:
equations.md— Key equations with variable definitions. Use for code verification.source.tex— Original LaTeX. Use when you need exact notation or the equation extraction missed something.summary.md— High-level overview focused on implementation (not literature review). Use for understanding the approach.tldr.md— Quick 2-3 sentence overview of the paper's contribution.figures/— Architecture diagrams, network structures, and result plots extracted from LaTeX source or PDF..pqa_papers/— Staged PDFs only (no markdown) so RAG backends don't index generated content.
Core commands
| Command | Purpose |
|---|---|
papi add <arxiv-id-or-url> |
Add a paper (downloads PDF + LaTeX, generates summary/equations/TL;DR) |
papi add --pdf file.pdf |
Add a local PDF or URL |
papi add --from-file list.json |
Import papers from a JSON list or text file |
papi list |
List papers (filter with --tag) |
papi search "query" |
Search across titles, tags, summaries, equations (--rg for grep-style, default uses ranked BM25 if indexed) |
papi index --backend search |
Build/update ranked search index (search.db) |
papi show <paper> --level eq |
Print equations (best for agent sessions) |
papi show <paper> --level tex |
Print LaTeX source |
papi show <paper> --level summary |
Print summary |
papi show <paper> --level tldr |
Print TL;DR |
papi export <papers...> --to ./dir |
Export context files into a repo (--level summary|equations|full) |
papi notes <paper> |
Open/print implementation notes |
papi regenerate <papers...> |
Regenerate summary/equations/tags/TL;DR |
papi remove <papers...> |
Remove papers |
papi ask "question" |
Cross-paper RAG query (requires PaperQA2 or LEANN) |
papi index |
Build/update the retrieval index |
papi tags |
List all tags |
papi path |
Print database location |
papi docs |
Print agent integration snippet (for CLAUDE.md/AGENTS.md) |
papi rebuild-index |
Rebuild index.json from on-disk paper directories (recovery) |
Run papi --help or papi <command> --help for full options.
Import/Export
Share your paper collection with others or back it up.
Export:
# Export full list to JSON
papi list --json > my_papers.json
# Export specific tag
papi list --tag "computer-vision" --json > cv_papers.json
Import:
# Import from JSON (preserves custom names and tags)
papi add --from-file my_papers.json
# Import from text file (one arXiv ID per line)
papi add --from-file paper_ids.txt --tags "imported"
# Import from BibTeX file (requires bibtexparser)
papi add --from-file papers.bib
# or install with BibTeX support:
# uv tool install paperpipe --with "paperpipe[bibtex]"
Semantic Scholar Support:
# Add papers from Semantic Scholar
papi add https://www.semanticscholar.org/paper/...
papi add 0123456789abcdef0123456789abcdef01234567 # S2 paper ID
Exact text search (fast, no LLM required):
papi search --rg "AdamW" # case-insensitive, literal string (default)
papi search --rg --case-sensitive "NeRF" # match exact case
papi search --rg --regex "Eq\\. [0-9]+" # regex mode (opt-in)
Ranked search (BM25 via SQLite FTS5, no LLM required):
papi index --backend search --search-rebuild # builds <paper_db>/search.db
papi search "surface reconstruction" # uses FTS if available (default)
papi search --no-fts "surface reconstruction" # force in-memory scan (disables FTS, uses fuzzy matching)
papi search --no-fts --exact "exact phrase" # force scan with exact matching only
Hybrid ranked+exact search:
papi search --hybrid "surface reconstruction"
papi search --hybrid --show-grep-hits "surface reconstruction"
What are FTS and BM25?
- FTS = Full-Text Search. Here it means SQLite’s FTS5 extension, which builds an inverted index so searches don’t have to rescan every file on every query.
- BM25 = Okapi BM25, a standard relevance-ranking function used by many search engines. It ranks results based on term frequency, inverse document frequency, and document length normalization.
References (external):
https://sqlite.org/fts5.html
https://en.wikipedia.org/wiki/Okapi_BM25
Glossary (RAG, embeddings, MCP, LiteLLM)
- RAG = retrieval‑augmented generation: retrieve relevant paper passages first, then generate an answer grounded in those passages.
- Embedding model = turns text into vectors for semantic search; changing it usually requires rebuilding an index.
- LiteLLM model id = the model string you pass to LiteLLM (provider/model routing), e.g.
gpt-4o,gemini/...,ollama/.... - MCP = Model Context Protocol: lets tools/agents call into paperpipe’s retrieval helpers (e.g. “retrieve chunks”) without copying PDFs into the chat.
- Staging dir (
.pqa_papers/) = PDF-only mirror used so RAG backends don’t index generated Markdown.
Config: default search mode
Set a default for papi search (CLI flags still win):
export PAPERPIPE_SEARCH_MODE=auto # auto|fts|scan|hybrid
Or in config.toml:
[search]
mode = "auto" # auto|fts|scan|hybrid
Agent integration
paperpipe is designed to work with coding agents. Install the skill and MCP servers:
papi install # installs skill + prompts + MCP for detected CLIs
# or be specific:
papi install skill --claude --codex --gemini
papi install mcp --claude --codex --gemini
After installation, your agent can:
- Use
/papito get paper context (skill) - Call MCP tools like
retrieve_chunksfor RAG retrieval - Verify code against paper equations
Custom prompts
| Prompt | Description |
|---|---|
/papi |
Route questions to the cheapest papi command |
/papi-init |
Add/update PaperPipe integration in your project's AGENTS.md/CLAUDE.md |
/verify-with-paper |
Verify code against paper equations |
/ground-with-paper |
Ground responses in paper excerpts |
/compare-papers |
Compare multiple papers for a decision |
/curate-paper-note |
Create a project note from paper excerpts |
For a ready-to-paste snippet for your repo's agent instructions, run papi docs or see AGENT_INTEGRATION.md.
What the agent sees
When you (or your agent) run papi show <paper> --level eq, you get structured output like:
## Equation 1: LoRA Update
$$h = W_0 x + \Delta W x = W_0 x + BA x$$
where:
- $W_0 \in \mathbb{R}^{d \times k}$: pretrained weight matrix (frozen)
- $B \in \mathbb{R}^{d \times r}$, $A \in \mathbb{R}^{r \times k}$: low-rank matrices
- $r \ll \min(d, k)$: the rank (typically 1-64)
This is what makes verification possible — the agent can compare your code symbol-by-symbol.
MCP server setup (manual)
MCP servers
paperpipe provides MCP servers for retrieval-only workflows:
- PaperQA2 retrieval: raw chunks + citations (via
paperqa_mcp) - LEANN search: fast semantic search over papers (via
leann_mcp)
MCP servers are configured automatically when you run papi install mcp. The install command creates the appropriate configuration files for your agent (Claude Code, Codex CLI, or Gemini CLI).
Installation:
# Install MCP servers for all supported agents (user scope)
papi install mcp
# Install for specific agents
papi install mcp --claude
papi install mcp --codex
papi install mcp --gemini
# Install repo-local MCP configs (Claude + Gemini) and Codex globally
papi install mcp --repo
# Customize embedding model
papi install mcp --embedding text-embedding-3-small
The MCP servers are automatically launched by your agent when needed. You don't need to manually start them.
MCP environment variables
| Variable | Default | Description |
|---|---|---|
PAPERPIPE_PQA_INDEX_DIR |
~/.paperpipe/.pqa_index |
Root directory for PaperQA2 indices |
PAPERPIPE_PQA_INDEX_NAME |
paperpipe_<embedding> |
Index name (subfolder under index dir) |
PAPERQA_EMBEDDING |
(from config) | Embedding model (must match index for PaperQA2) |
MCP tools
| Tool | Backend | Description |
|---|---|---|
retrieve_chunks |
PaperQA2 | Retrieve raw chunks + citations (no LLM answering) |
list_pqa_indexes |
PaperQA2 | List available PaperQA2 indices with embedding model metadata |
get_pqa_index_status |
PaperQA2 | Show index stats (files, failures) |
leann_search |
LEANN | Semantic search over papers (faster, simpler output) |
leann_list |
LEANN | List available LEANN indexes |
MCP usage
- Build indexes:
papi index --backend pqa --pqa-embedding text-embedding-3-small - In your agent:
leann_search()(fast) orretrieve_chunks()(with citations) - For PaperQA2: embedding model is automatically inferred from index metadata (or index name for backward compatibility)
RAG backends (papi ask)
paperpipe supports two RAG backends for cross-paper questions:
| Backend | Install | Best for |
|---|---|---|
| PaperQA2 | paperpipe[paperqa] |
Agentic synthesis with citations (cloud LLMs) |
| LEANN | paperpipe[leann] |
Local retrieval (Ollama) |
# PaperQA2 (default if installed)
papi ask "What regularization techniques do these papers use?"
# LEANN (local)
papi ask "..." --backend leann
The first query builds an index (cached under .pqa_index/ or .leann/). Use papi index to pre-build.
PaperQA2 configuration
Common options
| Flag | Description |
|---|---|
--pqa-llm MODEL |
LLM for answer generation (LiteLLM id) |
--pqa-summary-llm MODEL |
LLM for evidence summarization (often cheaper) |
--pqa-embedding MODEL |
Embedding model for text chunks |
--pqa-temperature FLOAT |
LLM temperature (0.0-1.0) |
--pqa-verbosity INT |
Logging level (0-3; 3 = log all LLM calls) |
--pqa-agent-type TEXT |
Agent type (e.g., fake for deterministic low-token retrieval) |
--pqa-answer-length TEXT |
Target answer length (e.g., "about 200 words") |
--pqa-evidence-k INT |
Number of evidence pieces to retrieve (default: 10) |
--pqa-max-sources INT |
Max sources to cite in answer (default: 5) |
--pqa-timeout FLOAT |
Agent timeout in seconds (default: 500) |
--pqa-concurrency INT |
Indexing concurrency (default: 1) |
--pqa-rebuild-index |
Force full index rebuild |
--pqa-retry-failed |
Retry previously failed documents |
--format evidence-blocks |
Output JSON with {answer, evidence[]} (requires PaperQA2 Python package) |
--pqa-raw |
Show raw PaperQA2 output (streaming logs + answer); disables papi ask output filtering (also enabled by global -v/--verbose) |
Any additional arguments are passed through to pqa (e.g., --agent.search_count 10).
Model combinations
Model combination examples
Indexing:
# API keys should be in env
export OPENAI_API_KEY=...
export ANTHROPIC_API_KEY=...
export GEMINI_API_KEY=...
export VOYAGE_API_KEY=...
# Ollama (local) + Ollama embeddings
papi index --backend pqa --pqa-llm ollama/olmo-3:7b --pqa-embedding ollama/nomic-embed-text
# GPT + OpenAI Embeddings
papi index --backend pqa --pqa-llm gpt-4.1 --pqa-summary-llm gpt-4.1-mini --pqa-embedding text-embedding-3-small
# Gemini + Google Embeddings
papi index --backend pqa --pqa-llm gemini/gemini-3-flash-preview --pqa-embedding gemini/gemini-embedding-001
# Claude + Voyage Embeddings
papi index --backend pqa --pqa-llm claude-sonnet-4-5 --pqa-summary-llm claude-haiku-4-5 --pqa-embedding voyage/voyage-3.5
Asking:
# Ollama (local)
papi ask "how is neus different from nerf?" --backend pqa --pqa-llm ollama/olmo-3:7b --pqa-embedding ollama/nomic-embed-text
# GPT
papi ask "how is neus different from nerf?" --backend pqa --pqa-llm gpt-4.1 --pqa-summary-llm gpt-4.1-mini --pqa-embedding text-embedding-3-small
# Gemini
papi ask "how is neus different from nerf?" --backend pqa --pqa-llm gemini/gemini-3-flash-preview --pqa-embedding gemini/gemini-embedding-001
# Claude
papi ask "how is neus different from nerf?" --backend pqa --pqa-llm claude-sonnet-4-5 --pqa-summary-llm claude-haiku-4-5 --pqa-embedding voyage/voyage-3.5
Embedding provider examples (indexing)
OpenAI
export OPENAI_API_KEY=...
papi index --backend pqa --pqa-embedding text-embedding-3-small
Gemini (native LiteLLM id)
export GEMINI_API_KEY=...
papi index --backend pqa --pqa-embedding gemini/gemini-embedding-001
Voyage (native LiteLLM id)
export VOYAGE_API_KEY=...
papi index --backend pqa --pqa-embedding voyage/voyage-3.5
OpenAI-compatible endpoints (advanced)
If you want to hit an OpenAI-compatible endpoint directly (instead of a native LiteLLM provider id), set
OPENAI_API_BASE and OPENAI_API_KEY and use an openai/... embedding id.
export OPENAI_API_BASE=https://api.voyageai.com/v1
export OPENAI_API_KEY="$VOYAGE_API_KEY"
papi index --backend pqa --pqa-embedding openai/voyage-3.5
Index/caching notes
- First run builds an index under
<paper_db>/.pqa_index/and stages PDFs under<paper_db>/.pqa_papers/. - Override index location with
PAPERPIPE_PQA_INDEX_DIR. - If you indexed wrong content (or changed embeddings), delete
.pqa_index/to force rebuild. - If PDFs failed indexing (recorded as
ERROR), re-run with--pqa-retry-failedor--pqa-rebuild-index. - By default,
papi askuses--settings defaultto avoid stale user settings; pass-s/--settings <name>to override. - If Pillow is not installed,
papi askforces--parsing.multimodal OFF; pass your own--parsing...args to override.
LEANN configuration
Common options
papi ask "..." --backend leann --leann-provider ollama --leann-model qwen3:8b
papi ask "..." --backend leann --leann-host http://localhost:11434
papi ask "..." --backend leann --leann-top-k 12 --leann-complexity 64
Notes:
- If you use
--leann-provider anthropic, yourleanninstall must include theanthropicPython package (pip install anthropicin the same environment that runsleann). - You can pass through extra
leannCLI flags after--(useful for debugging), e.g.:papi -v ask "..." --backend leann -- ...
Model combinations
Model combination examples
Indexing:
# API keys should be in env
export OPENAI_API_KEY=...
export ANTHROPIC_API_KEY=...
export GEMINI_API_KEY=...
export VOYAGE_API_KEY=...
# Ollama (local) + Ollama embeddings
papi index --backend leann --leann-embedding-mode ollama --leann-embedding-model nomic-embed-text
# OpenAI + OpenAI embeddings
papi index --backend leann --leann-embedding-mode openai --leann-embedding-model text-embedding-3-small --leann-embedding-api-key $OPENAI_API_KEY
# Gemini + Gemini embeddings (OpenAI-compatible)
papi index --backend leann --leann-embedding-mode openai --leann-embedding-model gemini-embedding-001 --leann-embedding-api-base https://generativelanguage.googleapis.com/v1beta/openai/ --leann-embedding-api-key $GEMINI_API_KEY
# Voyage embeddings (OpenAI-compatible)
papi index --backend leann --leann-embedding-mode openai --leann-embedding-model voyage-3.5 --leann-embedding-api-base https://api.voyageai.com/v1 --leann-embedding-api-key $VOYAGE_API_KEY
Asking:
# Ollama (local)
papi ask "how is neus different from nerf?" --backend leann --leann-provider ollama --leann-model olmo-3:7b --leann-index papers_ollama_nomic-embed-text
# OpenAI
papi ask "how is neus different from nerf?" --backend leann --leann-provider openai --leann-model gpt-4.1 --leann-api-key $OPENAI_API_KEY --leann-index papers_openai_text-embedding-3-small
# Anthropic + Voyage embeddings
papi ask "how is neus different from nerf?" --backend leann --leann-provider anthropic --leann-model claude-sonnet-4-5 --leann-api-key $ANTHROPIC_API_KEY --leann-index papers_openai_voyage-3.5
# Gemini (OpenAI-compatible)
papi ask "how is neus different from nerf?" --backend leann --leann-provider openai --leann-model gemini-3-flash-preview --leann-api-base https://generativelanguage.googleapis.com/v1beta/openai/ --leann-api-key $GEMINI_API_KEY --leann-index papers_openai_gemini-embedding-001
Embedding provider examples
Note: For --leann-embedding-mode openai, LEANN defaults the API key to OPENAI_API_KEY unless you pass --leann-embedding-api-key.
# Ollama (local)
papi index --backend leann --leann-embedding-mode ollama --leann-embedding-model nomic-embed-text
# OpenAI
export OPENAI_API_KEY=...
papi index --backend leann --leann-embedding-mode openai --leann-embedding-model text-embedding-3-small --leann-embedding-api-key $OPENAI_API_KEY
# Gemini (OpenAI-compatible)
export GEMINI_API_KEY=...
papi index --backend leann --leann-embedding-mode openai --leann-embedding-model gemini-embedding-001 --leann-embedding-api-base https://generativelanguage.googleapis.com/v1beta/openai/ --leann-embedding-api-key $GEMINI_API_KEY
# Voyage (OpenAI-compatible)
export VOYAGE_API_KEY=...
papi index --backend leann --leann-embedding-mode openai --leann-embedding-model voyage-3.5 --leann-embedding-api-base https://api.voyageai.com/v1 --leann-embedding-api-key $VOYAGE_API_KEY
Gemini notes:
- May hit quota/rate limits (HTTP 429). Retry after suggested delay.
- Some LEANN versions batch too many inputs per request for Gemini (hard limit: 100 inputs/request) and fail with HTTP 400; update LEANN or reduce chunk counts (e.g., larger
--leann-doc-chunk-size).
Defaults
By default, paperpipe derives LEANN's defaults from your global [llm] / [embedding] model settings when they are
LEANN-compatible:
ollama/...→--llm ollama/--embedding-mode ollamagpt-*/text-embedding-*→--llm openai/--embedding-mode openaigemini/...→--llm openai(Gemini OpenAI-compatible endpoint)
For Gemini, paperpipe defaults --leann-api-base to https://generativelanguage.googleapis.com/v1beta/openai/ and uses
GEMINI_API_KEY/GOOGLE_API_KEY if set.
Note: LEANN's current CLI batches OpenAI-compatible embeddings in chunks of up to ~500-800 texts per request; Gemini's
embedding endpoint hard-limits batches to 100, so paperpipe does not auto-map gemini/... embeddings to LEANN by
default. Use PAPERPIPE_LEANN_EMBEDDING_* / [leann] to override (and expect to tune batch behavior upstream in LEANN).
Multiple indices
LEANN supports multiple index names under <paper_db>/.leann/indexes/.
By default, paperpipe auto-derives the LEANN index name from the embedding mode/model (similar to PaperQA2).
To disable and always use a single LEANN index named papers, set:
[leann]
index_by_embedding = false
or export PAPERPIPE_LEANN_INDEX_BY_EMBEDDING=0.
When enabled, the default LEANN index name becomes papers_<mode>_<model> (with / and : replaced by _).
If model ids are not recognized as compatible, it falls back to ollama with olmo-3:7b (LLM) and nomic-embed-text
(embeddings).
Override via config.toml:
[leann]
llm_provider = "ollama"
llm_model = "qwen3:8b"
embedding_model = "nomic-embed-text"
embedding_mode = "ollama"
Or env vars: PAPERPIPE_LEANN_LLM_PROVIDER, PAPERPIPE_LEANN_LLM_MODEL, PAPERPIPE_LEANN_EMBEDDING_MODEL, PAPERPIPE_LEANN_EMBEDDING_MODE.
Index builds
papi index --backend leann
# Override common LEANN build knobs (maps to `leann build ...`):
papi index --backend leann --leann-embedding-mode ollama --leann-embedding-model nomic-embed-text
papi index --backend leann --leann-embedding-mode ollama --leann-embedding-host http://localhost:11434
papi index --backend leann --leann-doc-chunk-size 350 --leann-doc-chunk-overlap 128
By default, papi ask --backend leann auto-builds the index if missing (disable with --leann-no-auto-index).
LLM configuration
paperpipe uses LLMs for generating summaries, extracting equations, and tagging. Without an LLM, it falls back to regex extraction and metadata-based summaries.
# Set your API key (pick one)
export GEMINI_API_KEY=... # default provider
export OPENAI_API_KEY=...
export ANTHROPIC_API_KEY=...
export VOYAGE_API_KEY=... # for Voyage embeddings (recommended with Claude)
export OPENROUTER_API_KEY=... # 200+ models
# Override the default model
export PAPERPIPE_LLM_MODEL=gpt-4o
export PAPERPIPE_LLM_TEMPERATURE=0.3 # default: 0.3
Local-only via Ollama
export PAPERPIPE_LLM_MODEL=ollama/qwen3:8b
export PAPERPIPE_EMBEDDING_MODEL=ollama/nomic-embed-text
# Either env var name works (paperpipe normalizes both):
export OLLAMA_HOST=http://localhost:11434
# export OLLAMA_API_BASE=http://localhost:11434
Check which models work with your keys:
papi models # probe default models for your configured keys
papi models latest # probe latest models (gpt-4o, gemini-2.5, claude-sonnet-4-5)
papi models last-gen # probe previous generation
papi models all # probe broader superset
papi models --verbose # show underlying provider errors
Tagging
Papers are auto-tagged from:
- arXiv categories (cs.CV → computer-vision)
- LLM-generated semantic tags
- Your
--tagsflag
papi add 1706.03762 --tags my-project,priority
papi list --tag attention
Non-arXiv papers
papi add --pdf ./paper.pdf
papi add --pdf "https://example.com/paper.pdf" --tags siggraph
Configuration file
For persistent settings, create ~/.paperpipe/config.toml (override location with PAPERPIPE_CONFIG_PATH):
[llm]
model = "gemini/gemini-2.5-flash"
temperature = 0.3
[embedding]
model = "gemini/gemini-embedding-001"
[paperqa]
settings = "default"
index_dir = "~/.paperpipe/.pqa_index"
summary_llm = "gpt-4o-mini"
enrichment_llm = "gpt-4o-mini"
# Optional: override LEANN separately (otherwise it follows [llm]/[embedding] for openai/ollama model ids)
[leann]
llm_provider = "ollama"
llm_model = "qwen3:8b"
embedding_model = "nomic-embed-text"
embedding_mode = "ollama"
[tags.aliases]
cv = "computer-vision"
nlp = "natural-language-processing"
Precedence: CLI flags > env vars > config.toml > built-in defaults.
Development
git clone https://github.com/hummat/paperpipe && cd paperpipe
pip install -e ".[dev]"
make check # format + lint + typecheck + test
Release (maintainers)
This repo publishes to PyPI when a GitHub Release is published (see .github/workflows/publish.yml).
# Bump version in pyproject.toml, then:
make release
Credits
- PaperQA2 by Future House — RAG backend. Skarlinski et al., "Language Agents Achieve Superhuman Synthesis of Scientific Knowledge", 2024. arXiv:2409.13740
- LEANN — (local) RAG backend. Wang et al., "LEANN: A Low-Storage Vector Index", 2025. arXiv:2506.08276
License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file paperpipe-1.3.0.tar.gz.
File metadata
- Download URL: paperpipe-1.3.0.tar.gz
- Upload date:
- Size: 136.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
77e76011a9934d0d00e608639491099e46906f91f2ee5bb422b148a2527741f0
|
|
| MD5 |
3690f5ee7bd89ea15f95ffeac68fd381
|
|
| BLAKE2b-256 |
bd0f60d0183f91c022e68e69f52fa449051a3e658fde61acb2ec9263022ae15f
|
Provenance
The following attestation bundles were made for paperpipe-1.3.0.tar.gz:
Publisher:
publish.yml on hummat/paperpipe
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
paperpipe-1.3.0.tar.gz -
Subject digest:
77e76011a9934d0d00e608639491099e46906f91f2ee5bb422b148a2527741f0 - Sigstore transparency entry: 836411009
- Sigstore integration time:
-
Permalink:
hummat/paperpipe@d72383fc7080afb86a87d46e7b21921c8fcaec51 -
Branch / Tag:
refs/tags/v1.3.0 - Owner: https://github.com/hummat
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@d72383fc7080afb86a87d46e7b21921c8fcaec51 -
Trigger Event:
release
-
Statement type:
File details
Details for the file paperpipe-1.3.0-py3-none-any.whl.
File metadata
- Download URL: paperpipe-1.3.0-py3-none-any.whl
- Upload date:
- Size: 165.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
cf30611baf2ff3484b526019bf012d9b81af020cb2d29bbdddfd3716034c8593
|
|
| MD5 |
d0fe0f54ea628b075efa5f6baf22578e
|
|
| BLAKE2b-256 |
369863c68039634dfb9cc08cd31120e75256e2978e3ef5935f04f8aedec5dcea
|
Provenance
The following attestation bundles were made for paperpipe-1.3.0-py3-none-any.whl:
Publisher:
publish.yml on hummat/paperpipe
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
paperpipe-1.3.0-py3-none-any.whl -
Subject digest:
cf30611baf2ff3484b526019bf012d9b81af020cb2d29bbdddfd3716034c8593 - Sigstore transparency entry: 836411014
- Sigstore integration time:
-
Permalink:
hummat/paperpipe@d72383fc7080afb86a87d46e7b21921c8fcaec51 -
Branch / Tag:
refs/tags/v1.3.0 - Owner: https://github.com/hummat
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@d72383fc7080afb86a87d46e7b21921c8fcaec51 -
Trigger Event:
release
-
Statement type: