Skip to main content

Library to create and interogate local cache for Project Gutenberg

Project description



.. image::
:align: center
:width: 100%

This package makes filtering and getting information from `Project
Gutenberg <>`_ easier from python.

It's target audience is machine learning guys that need data for their project,
but may be freely used by anybody.

The package:

- Generates a local cache (of all gutenberg informations) that you can interogate to get book ids. The Local cache may be sqlite (default) or mongodb (for wich you need to have installed the pymongodb packet)

- Downloads and cleans raw text from gutenberg books

The package has been tested with Python 2.7 on both Windows and Linux
It is faster, smaller and less third-party intensive alternative to

About development:


.. sourcecode :: sh

pip install gutenbergpy

or just install it from source (it's all just python code)

.. sourcecode :: sh

git clone
python install


Downloading a text

.. sourcecode :: python

import gutenbergpy.textget
#this gets a book by its gutenberg id
raw_book = gutenbergpy.textget.get_text_by_id(1000)
print raw_book
#this strips the headers from the book
clean_book = gutenbergpy.textget.strip_headers(raw_book)
print clean_book

Query the cache
To do this you first need to create the cache (this is a one time thing per os, until you decide to redo it)

.. sourcecode :: python

from gutenbergpy.gutenbergcache import GutenbergCache
#for sqlite
#for mongodb

for debugging/better control you have these boolean options on create

- *refresh* deletes the old cache
- *download* property downloads the rdf file from the gutenberg project
- *unpack* unpacks it
- *parse* parses it in memory
- *cache* writes the cache

.. sourcecode :: python

GutenbergCache.create(refresh=True, download=True, unpack=True, parse=True, cache=True, deleteTemp=True)

for even better control you may set the GutenbergCacheSettings
- *CacheFilename*
- *CacheUnpackDir*
- *CacheArchiveName*
- *ProgressBarMaxLength*
- *CacheRDFDownloadLink*
- *TextFilesCacheFolder*
- *MongoDBCacheServer*
.. sourcecode :: python

GutenbergCacheSettings.set( CacheFilename="", CacheUnpackDir="",
CacheArchiveName="", ProgressBarMaxLength="", CacheRDFDownloadLink="", TextFilesCacheFolder="", MongoDBCacheServer="")

After doing a create you need to wait, it will be over in about 5 minutes depending on your internet speed and computer power
(On a i7 with gigabit connection and ssd it finishes in about 1 minute)

Get the cache

.. sourcecode :: python
#for mongodb
cache = GutenbergCache.get_cache(GutenbergCacheTypes.CACHE_TYPE_MONGODB)
#for sqlite
cache = GutenbergCache.get_cache()

Now you can do queries

Get the book Gutenberg unique indices by using this query function

Standard query fields:
- languages
- authors
- types
- titles
- subjects
- publishers
- bookshelves
- downloadtype

.. sourcecode :: python

print cache.query(downloadtype=['application/plain','text/plain','text/html; charset=utf-8'])

Or do a native query on the sqlite database

.. sourcecode :: python
cache.native_query("SELECT * FROM books")

For SQLITE custom queries take a look at the SQLITE database scheme:

.. image::
:alt: SQLITE database scheme
:width: 100%
:align: center

For MongoDB queries you have all the books collection. Each book with the following fields:

- book(publisher, rights, language, book_shelf, gutenberg_book_id, date_issued, num_downloads, titles, subjects, authors, files ,type)

Project details

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Files for GutenbergPy, version 0.2.0
Filename, size & hash File type Python version Upload date (25.4 kB) View hashes Source None

Supported by

Elastic Elastic Search Pingdom Pingdom Monitoring Google Google BigQuery Sentry Sentry Error logging AWS AWS Cloud computing DataDog DataDog Monitoring Fastly Fastly CDN SignalFx SignalFx Supporter DigiCert DigiCert EV certificate StatusPage StatusPage Status page