Skip to main content

Iterates Django querysets by chunks, saving memory and allows to start faster.

Project description

# Django chunked iterator

Django provides a simple way to make complex queries.
Unfortunately, another side of the coin is
high memory consumption for really huge datasets.

The thing is that when you are getting an object from
[QuerySet](https://docs.djangoproject.com/en/2.1/ref/models/querysets/),
Django makes a query for all objects and makes model instances
for *all returned rows in memory*.
Even if you only need one object at a time.
Memory also will be consumed for holding result rows
in the database and in Python's database driver.

```python
for e in Entry.objects.all():
print(e.headline)
```

There is a way to improve it:
[iterator](https://docs.djangoproject.com/en/2.1/ref/models/querysets/#iterator).

```python
for e in Entry.objects.iterator():
print(e.headline)
```

In this way, Django will construct model instances
on the fly only for current iteration.
Depending on your database and settings,
Django can also get all rows from database in one query,
or it can use server-side cursors to get rows by chunks.

In the latter case (with server-side cursors),
only limited amount of memory will be consumed
in the database and in Python's database driver.
But this only works with certain databases
and without connection poolers (like [pgbouncer](https://pgbouncer.github.io)).
There is no way for your code to be sure that
the memory-efficient method is used.

## Design

This chunked iterator takes queryset and makes serial queries
returning fixed number of rows or model instances.
This allows iterating really huge amount of rows
with fixed memory consumption on the database,
Python's driver, and application layers.
As a side effect, the first portion of rows is returned much quicker,
which allows starting processing in parallel in some cases.

There is only one limitation: the model should have a unique field
which will be used for sorting and pagination.
For most cases, this is the primary key, but you can use other fields.

## Installing

```bash
$ pip install django-chunked-iterator
```

## Using

#### Iterate over `QuerySet` by chunks

```python
from django_chunked_iterator import batch_iterator

for entries in batch_iterator(Entry.objects.all()):
for e in entries:
print(e.headline)
```

#### Iterate over `QuerySet` by items

```python
from django_chunked_iterator import iterator

for e in iterator(Entry.objects.all()):
print(e.headline)
```

#### Limit number of returned rows

WRONG!

```python
for e in iterator(Entry.objects.all()[:10000]):
print(e.headline)
AssertionError: Cannot reorder a query once a slice has been taken.
```

Right:

```python
for e in iterator(Entry.objects.all(), limit=10000):
print(e.headline)
```

#### Change batch size

The smaller batch size, the faster first item is returned,
the larger overhead for additional queries.
Optimal values from 100 to 1000.

```python
for e in iterator(Entry.objects.all(), batch_size=150):
print(e.headline)
```

#### Change ordering

Returns items in reverse creation order.
`created` field **should** have `uniquie=True`
(this is not hard when datetime has microseconds accuracy).

```python
for e in iterator(Entry.objects.all(), order_by='-created'):
print(e.headline)
```


## Testing

```bash
$ pip install -r ./requirements.txt
$ ./test_project/manage.py test -v 2 --with-coverage --cover-package=django_chunked_iterator
```

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

django-chunked-iterator-0.6.1.tar.gz (3.3 kB view details)

Uploaded Source

File details

Details for the file django-chunked-iterator-0.6.1.tar.gz.

File metadata

File hashes

Hashes for django-chunked-iterator-0.6.1.tar.gz
Algorithm Hash digest
SHA256 89b73870f3dce2c688fdbec2eb30bacb7dedc5d138a7cb36d2f47baa58411015
MD5 57ad51ca87a4852ff97c58e04ecda45b
BLAKE2b-256 c074c7091a5bcf21283c67cde1b3c7f36406355c4cc697e90699341e0fb32545

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page