A service for handling deferred and periodic tasks on beanstalk worker machines

# django-beanstalk-worker

An app for handling deferred and periodic tasks on beanstalk worker machines

## Overview

The core of this app is the task decorator.

When a decorated task function is called instead of running immediately an SQS message is queued. When unqueued by a worker that message will cause the original function to be called.

The arguments may be any combination of JSON serializable types, datetimes and decimals, including keyword arguments. Any other argument types will cause an error.

Queuing of SQS messages happens in transaction on_commit so if you exception out of a transaction any tasks called in that transaction will not be run.

Task functions can also be invoked by cron or a management command as needed.

## Installation and Setup

### 1: Installed Apps

Add beanstalk_worker to installed_apps.

### 2: Environment Variables

In the Software section of the Beanstalk configuration of your worker environment add the environment variable WORKER and set it to 1

### 3: Django Settings

BEANSTALK_WORKER = bool(os.environ.get("WORKER", False))
BEANSTALK_SQS_URL = <SQS queue URL>
BEANSTALK_SQS_REGION = <Amazon region>


You can find the SQS queue URL in the Worker section of the Beanstalk environment configuration.

For test and development you can omit the SQS settings and should set BEANSTALK_TASK_SERVICE = "beanstalk_worker.services.FakeTaskService"

### 4: URLs's

if settings.BEANSTALK_WORKER:


This will add the URLS /tasks/task/ and /tasks/cron. You can move the base of these URL's if you do so other instructions on this page will need updating appropriately.

DO NOT include these URL's in a production web server, only the worker.

For test and development you won't have seperate web and worker machines so always include the URL's.

### 5: Beanstalk Worker Configuration

In the Worker section of the Beanstalk configuration of your worker environment set HTTP path to /tasks/task/ and MIME type to application/json

## Use

from beanstalk_worker import task

print(message)


my_task("hello world")


This will queue an SQS message instructing a worker to run the actual function.

The arguments may be any combination of JSON serializable types, datetimes and decimals, including keyword arguments

### Call the task from CRON

- name: "my_project.my_app.tasks.my_task"
schedule: "0 0 * * *"


my_project.my_app.tasks.my_task should be replaced with the fully qualified name of your task function.

Arguments are not currently supported for cron.

### Call the task from the command line

./manage.py run_task my_project.my_app.tasks task "['hello world']"


my_project.my_app.tasks my_task should be replaced with the fully qualified name of your task function, also note the space between module name and function name.

## Development, Testing and the FakeTaskServer

The FakeTaskServer will internally queue tasks but will not run them unless instructed to.

### settings.DEBUG

When DEBUG = True is set in the Django settings an additional URL is exposed at /tasks/run_all/ if you poke this URL the FakeTaskService will run all queued tasks. This can be handy for local development.

### Test support

In tests you can acquire the running task service instance with from beanstalk_worker import task_service. This class has two helpers clear whcih will discard all queued tasks and run_all which will immediately run all queued tasks.

### settings.BEANSTALK_WORKER

While running a task the FakeTaskServer will patch settings.BEANSTALK_WORKER to True. This lets you assert settings.BEANSTALK_WORKER in code that should only ever be run on a worker.

## Project details

Uploaded source
Uploaded py3