A Celery Beat Scheduler using Redis for persistent storage
Project description
RedBeat is a Celery Beat Scheduler that stores the scheduled tasks and runtime metadata in Redis.
Why RedBeat
Dynamic live task creation and modification, without lengthy downtime
Externally manage tasks from any language with Redis bindings
Shared data store; Beat isn’t tied to a single drive or machine
Fast startup even with a large task count
Prevent accidentally running multiple Beat servers
Getting Started
Install with pip:
pip install celery-redbeat
Configure RedBeat settings in your Celery configuration file:
redbeat_redis_url = "redis://localhost:6379/1"
Then specify the scheduler when running Celery Beat:
celery beat -S redbeat.RedBeatScheduler
RedBeat uses a distributed lock to prevent multiple instances running. To disable this feature, set:
redbeat_lock_key = None
Configuration
You can add any of the following parameters to your Celery configuration (see Celery 3.x compatible configuration value names in below).
redbeat_redis_url
URL to redis server used to store the schedule, defaults to value of broker_url.
redbeat_key_prefix
A prefix for all keys created by RedBeat, defaults to 'redbeat'.
redbeat_lock_key
Key used to ensure only a single beat instance runs at a time, defaults to '<redbeat_key_prefix>:lock'.
redbeat_lock_timeout
Unless refreshed the lock will expire after this time, in seconds.
Defaults to five times of the default scheduler’s loop interval (300 seconds), so 1500 seconds (25 minutes).
See the beat_max_loop_interval Celery docs about for more information.
Celery 3.x config names
Here are the old names of the configuration values for use with Celery 3.x.
Celery 4.x |
Celery 3.x |
---|---|
redbeat_redis_url |
REDBEAT_REDIS_URL |
redbeat_key_prefix |
REDBEAT_KEY_PREFIX |
redbeat_lock_key |
REDBEAT_LOCK_KEY |
redbeat_lock_timeout |
REDBEAT_LOCK_TIMEOUT |
Design
At its core RedBeat uses a Sorted Set to store the schedule as a priority queue. It stores task details using a hash key with the task definition and metadata.
The schedule set contains the task keys sorted by the next scheduled run time.
For each tick of Beat
get list of due keys and due next tick
retrieve definitions and metadata for all keys from previous step
update task metadata and reschedule with next run time of task
call due tasks using async_apply
calculate time to sleep until start of next tick using remaining tasks
Creating Tasks
You can use Celery’s usual way to define static tasks or you can insert tasks directly into Redis. The config options is called beat_schedule, e.g.:
app.conf.beat_schedule = {
'add-every-30-seconds': {
'task': 'tasks.add',
'schedule': 30.0,
'args': (16, 16)
},
}
On Celery 3.x the config option was called CELERYBEAT_SCHEDULE.
The easiest way to insert tasks from Python is it use RedBeatSchedulerEntry():
interval = celey.schedulers.schdule(run_every=60) # seconds entry = RedBeatSchedulerEntry('task-name', 'tasks.some_task', interval, args=['arg1', 2]) entry.save()
Alternatively, you can insert directly into Redis by creating a new hash with a key of <redbeat_key_prefix>:task-name. It should contain a single key definition which is a JSON blob with the task details.
Interval
An interval task is defined with the JSON like:
{ "name" : "interval example", "task" : "tasks.every_5_seconds", "schedule": { "__type__": "interval", "every" : 5, # seconds "relative": false, # optional }, "args" : [ # optional "param1", "param2" ], "kwargs" : { # optional "max_targets" : 100 }, "enabled" : true, # optional }
Crontab
An crontab task is defined with the JSON like:
{ "name" : "crontab example", "task" : "tasks.daily", "schedule": { "__type__": "crontab", "minute" : "5", # optional, defaults to * "hour" : "*", # optional, defaults to * "day_of_week" : "monday", # optional, defaults to * "day_of_month" : "*/7", # optional, defaults to * "month_of_year" : "[1-12]", # optional, defaults to * }, "args" : [ # optional "param1", "param2" ], "kwargs" : { # optional "max_targets" : 100 }, "enabled" : true, # optional }
Scheduling
Assuming your redbeat_key_prefix config values is set to ‘redbeat:’ (default) you will also need to insert the new task into the schedule with:
zadd redbeat::schedule 0 new-task-name
The score is the next time the task should run formatted as a UNIX timestamp.
Metadata
Applications may also want to manipulate the task metadata to have more control over when a task runs. The meta key contains a JSON blob as follows:
{ 'last_run_at': { '__type__': 'datetime', 'year': 2015, 'month': 12, 'day': 29, 'hour': 16, 'minute': 45, 'microsecond': 231 }, 'total_run_count'; 23 }
For instance by default `last_run_at` corresponds to when Beat dispatched the task, but depending on queue latency it might not run immediately, but the application could update the metadata with the actual run time, allowing intervals to be relative to last execution rather than last dispatch.
Development
RedBeat is available on GitHub
Once you have the source you can run the tests with the following commands:
pip install -r requirements.dev.txt py.test tests
You can also quickly fire up a sample Beat instance with:
celery beat --config exampleconf
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.