Skip to main content

Tools for working with the Jeffrey Epstein documents released in November 2025.

Project description

Color Highlighted Epstein Emails and Text Messages

joi

Usage

Installation

Use poetry install for easiest time installing. pip install epstein-files should also work, though pipx install epstein-files is usually better.

Then there's two options as far as the data:

  1. To work with the data set included in this repo copy the pickled data file into place: cp ./the_epstein_files.pkl.gz ./the_epstein_files.local.pkl.gz
  2. To parse your own files:
    1. Requires you have a local copy of the OCR text files from the House Oversight document release in a directory /path/to/epstein/ocr_txt_files. You can download those OCR text files from the Congressional Google Drive folder (make sure you grab both the 001/ and 002/ folders).
    2. (Optional) If you want to work with the documents released by DOJ on January 30th 2026 you'll need to also download some of the PDF files from the DOJ site (they're in the "Epstein Files Transparency Act" section). You don't need them all, just the ones you want to look at and make ASCII art with. But you will need to get the OCR text out o them somehow. I use pdfalyzer. IMPORTANT if

Command Line Tools

You need to set the EPSTEIN_DOCS_DIR environment variable with the path to the folder of files you just downloaded when running. You can either create a .env file modeled on .env.example (which will set it permanently) or you can run with:

EPSTEIN_DOCS_DIR=path/to/source_data/ epstein_generate --help

To work with the January 2026 DOJ documents you'll also need to set the EPSTEIN_DOJ_TXTS_20260130_DIR env var to point at folders full of OCR extracted texts from the raw DOJ PDFs. If you have the PDFs but not the text files there's a script that can help you take care of that (it launches pdfalyzer on PDFs it finds in the hierarchy).

EPSTEIN_DOCS_DIR=path/to/source_data/ EPSTEIN_DOJ_TXTS_20260130_DIR=/path/to/doj/files/ epstein_generate --help

NOTE: In order to get the generated links to the DOJ site and Jmail to work correctly you will need to sort the PDFs into the same datasets they are found in on the DOJ's website. You should have folders like this:

└── source_data/
    ├── DataSet 1
    ├── DataSet 2
    ├── DataSet 3
    ├── DataSet 4
    ├── DataSet 5
    ├── DataSet 6
    ├── DataSet 7
    ├── DataSet 8
    ├── DataSet 9
    ├── DataSet 10
    ├── DataSet 11
    └── DataSet 12

Within the DataSet N folders the PDFs can be sorted however you want (the folders will be recursively scanned for the pattern **/*.pdf).

Doing Things

All the tools that come with the package require EPSTEIN_DOCS_DIR to be set. These are the available tools:

# Generate color highlighted texts/emails/other files
epstein_generate

# Search for a string:
epstein_grep Bannon
# Or a regex:
epstein_grep '\bSteve\s*Bannon|Jeffrey\s*Epstein\b'

# Show a file with color highlighting of keywords:
epstein_show 030999
# Show both the highlighted and raw versions of the file:
epstein_show --raw 030999
# The full filename is also accepted:
epstein_show HOUSE_OVERSIGHT_030999

# Count words used by Epstein and Bannon
epstein_show --output-word-count --name 'Jeffrey Epstein' --name 'Steve Bannon'

# Diff two epstein files after all the cleanup (stripping BOMs, matching newline chars, etc):
epstein_diff 030999 020442

The first time you run anything it will take a few minutes to fix all the janky OCR text, attribute the redacted emails, etc. After that things will be quick.

The commands used to build the various sites that are deployed on Github Pages can be found in deploy.sh.

Run epstein_generate --help for command line option assistance.

Optional: There are a handful of emails that I extracted from the legal filings they were contained in. If you want to include these files in your local analysis you'll need to copy those files from the repo into your local document directory. Something like:

cp ./emails_extracted_from_legal_filings/*.txt "$EPSTEIN_DOCS_DIR"

As A Library

from epstein_files.epstein_files import EpsteinFiles
epstein_files = EpsteinFiles.get_files()

# All files
for document in epstein_files.documents:
    do_stuff(document)

# Emails
for email in epstein_files.emails:
    do_stuff(email)

# iMessage Logs
for imessage_log in epstein_files.imessage_logs:
    do_stuff(imessage_log)

# Other Files
for file in epstein_files.other_files:
    do_stuff(file)

Everyone Who Sent or Received an Email in the November Document Dump

emails

TODO List

See TODO.md.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

epstein_files-1.8.8.tar.gz (247.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

epstein_files-1.8.8-py3-none-any.whl (272.7 kB view details)

Uploaded Python 3

File details

Details for the file epstein_files-1.8.8.tar.gz.

File metadata

  • Download URL: epstein_files-1.8.8.tar.gz
  • Upload date:
  • Size: 247.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.6.1 CPython/3.11.11 Darwin/22.6.0

File hashes

Hashes for epstein_files-1.8.8.tar.gz
Algorithm Hash digest
SHA256 80de4b2ec19c1654aa6440dd57d0264eaec2742d4a3ebd908fae23839155eacb
MD5 66f6f274ffe454ab82e095b597c4306a
BLAKE2b-256 5f445da50f595b217aeb88f711a972f34d48785dc37a3fd97d5c4b42542751e7

See more details on using hashes here.

File details

Details for the file epstein_files-1.8.8-py3-none-any.whl.

File metadata

  • Download URL: epstein_files-1.8.8-py3-none-any.whl
  • Upload date:
  • Size: 272.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.6.1 CPython/3.11.11 Darwin/22.6.0

File hashes

Hashes for epstein_files-1.8.8-py3-none-any.whl
Algorithm Hash digest
SHA256 22420a7575e32163bbd206552ccaa68fe825f18b389a18cad486e1dee12c719c
MD5 32aa0f46bfd3b4f95b95dce725d6013b
BLAKE2b-256 0c4968a92183f559e1375f69ad5463b3b89603000fd8d60acc95af09dae68b13

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page