Skip to main content

Lazy reading of file objects for efficient batch processing

Project description

lazyreader is a Python module for doing lazy reading of file objects.

The Python standard library lets you read a file a line-at-a-time, saving you from loading the entire file into memory. For example:

with open('large_file.txt') as f:
    for line in f:
        print(line)

lazyreader lets you do the same thing, but with an arbitrary delimiter, and for any object that presents a .read() method. For example:

from lazyreader import lazyread

with open('large_file.txt') as f:
    for doc in lazyread(f, delimiter=';'):
        print(doc)

This is a snippet of code I spun out from the Wellcome Digital Platform. We have large XML and JSON files stored in S3 – sometimes multiple GBs – but each file is really a series of “documents”, separated by known delimiters. Downloading and parsing the entire file would be prohibitively expensive, but lazyreader allows us to hold just a single document in memory at a time.

Installation

lazyreader is available from PyPI:

$ pip install lazyreader

Examples

If we have a file stored locally, we can open it and split based on any choice of delimiter. For example, if we had a text file in which record were separated by commas:

with open('lots_of_records.txt') as f:
    for doc in lazyread(f, delimiter=','):
        print(doc)

Another example: we have a file stored in Amazon S3, and we’d like to read it line-by-line. The boto3 API gives us a file object for reading from S3:

import boto3

client = boto3.client('s3')
s3_object = client.get_object(Bucket='example-bucket', Key='words.txt')
body = s3_object['Body']

for doc in lazyread(body, delimiter=b'\n'):
    print(doc)

(This is the use case for which this code was originally written.)

One more example: we’re fetching an HTML page, and want to read lines separated by <br> in the underlying HTML. Like so:

import urllib.request

with urllib.request.urlopen('https://example.org/') as f:
    for doc in lazyread(f, delimiter=b'<br>'):
        print(doc)

Advanced usage

lazyread() returns a generator, which you can wrap to build a pipeline of generators which do processing on the data.

First example: we have a file which contains a list of JSON objects, one per line. (This is the format of output files from elasticdump.) What the caller really needs is Python dictionaries, not JSON strings. We can wrap lazyread() like so:

import json

def lazyjson(f, delimiter=b'\n'):
    for doc in lazyread(f, delimiter=delimiter):

        # Ignore empty lines, e.g. the last line in a file
        if not doc.strip():
            continue

        yield json.loads(doc)

Another example: we want to parse a large XML file, but not load it all into memory at once. We can write the following wrapper:

from lxml import etree

def lazyxmlstrings(f, opening_tag, closing_tag):
    for doc in lazyread(f, delimiter=closing_tag):
        if opening_tag not in doc:
            continue

        # We want complete XML blocks, so look for the opening tag and
        # just return its contents
        block = doc.split(opening_tag)[-1]
        yield opening_tag + block

def lazyxml(f, opening_tag, closing_tag):
    for xml_string in lazyxmlstrings(f, opening_tag, closing_tag):
         yield etree.fromstring(xml_string)

We use both of these wrappers at Wellcome to do efficient processing of large files that are kept in Amazon S3.

Isn’t this a bit simple to be a module?

Maybe. There are recipes on Stack Overflow that do very similar, but I find it useful to have in a standalone module.

And it’s not completely trivial – at least, not for me. I made two mistakes when I first wrote this:

  • I was hard-coding the initial running string as

    running = b''

    That only works if your file object is returning bytestrings. If it’s returning Unicode strings, you get a TypeError (can’t concat bytes to str) when it first tries to read from the file. String types are important!

  • After I’d read another 1024 characters from the file, I checked for the delimiter like so:

    running += new_data
    if delimiter in running:
        curr, running = running.split(delimiter)
        yield curr + delimiter

    For my initial use case, individual documents were much bigger than 1024 characters, so the new data would never contain multiple delimiters. But with smaller documents, you might get multiple delimiters in one read, and then unpacking the result of .split() would throw a ValueError. So now the code correctly checks and handles the case where a single read includes more than one delimiter.

Now it’s encoded and tested in a module, I don’t have to worry about making the same mistakes again.

License

MIT.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lazyreader-1.0.1.tar.gz (4.5 kB view details)

Uploaded Source

Built Distribution

lazyreader-1.0.1-py2-none-any.whl (7.2 kB view details)

Uploaded Python 2

File details

Details for the file lazyreader-1.0.1.tar.gz.

File metadata

  • Download URL: lazyreader-1.0.1.tar.gz
  • Upload date:
  • Size: 4.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No

File hashes

Hashes for lazyreader-1.0.1.tar.gz
Algorithm Hash digest
SHA256 12065c91f2ad743eecf59c0df2dba4ecc7cc015aa177dee2be1c76b03e3907d8
MD5 184799baf8d848a3dfbeb10a93c5cd9f
BLAKE2b-256 d64b6b310bced42a4777819d94d1734a6da9978fcde2c297fccb07a3bb5eb3f2

See more details on using hashes here.

File details

Details for the file lazyreader-1.0.1-py2-none-any.whl.

File metadata

File hashes

Hashes for lazyreader-1.0.1-py2-none-any.whl
Algorithm Hash digest
SHA256 629dceca1a9e2b9006170936b34cb58e7dbe163e65fca184286cf44c6e50967b
MD5 3e115e5383fa9111911049d102a8d830
BLAKE2b-256 5c7fa17e8eb62cc5cbd6aa6cfa0faca6d360eca6119433623a5bb4d58d58b079

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page