No project description provided
Project description
Abstract Block Dumper
This package provides a simplified framework for creating block processing tasks in Django applications. Define tasks with lambda conditions using the @block_task decorator and run them asynchronously with Celery.
Implementation Details
General Workflow:
Register functions -> detect new blocks -> evaluate conditions -> send to Celery -> execute -> track results -> handle retries.
WorkflowSteps
- Register
- Functions are automatically discovered when the scheduler starts
- Functions must be located in installed apps in tasks.py or block_tasks.py
- Functions marked with @block_task decorators are stored in memory registry
- Detect Blocks
- Scheduler is running by management command block_tasks
- Scheduler polls blockchain, finds new blocks, and batches them
- Plan Tasks
- For each block, lambda conditions are evaluated against registered functions
- Tasks are created for matching conditions (with optional multiple argument sets)
-
Queue Tasks are sent to Celery with queue and timeout settings from celery_kwargs
-
Execute Celery runs the function with block info, capturing results and errors
-
Track Task attempts are stored in TaskAttempt model with retry logic and state tracking
Prerequisites
- Django
- Celery
- Redis (for Celery broker and result backend)
- PostgreSQL (recommended for production)
Installation
- Install the package:
pip install abstract_block_dumper
- Add to your Django
INSTALLED_APPS:
INSTALLED_APPS = [
# ... other apps
'abstract_block_dumper',
]
- Run migrations:
python manage.py migrate
Usage
1. Define Block Processing Tasks
Create block processing tasks in tasks.py or block_tasks.py file inside any of your installed Django apps.
2. Use Decorators to Register Tasks
- Use
@block_taskwith lambda conditions to create custom block processing tasks
3. Start the Block Scheduler
Run the scheduler to start processing blocks:
$ python manage.py block_tasks
This command will:
- Automatically discover and register all decorated functions
- Start polling the blockchain for new blocks
- Schedule tasks based on your lambda conditions
4. Start Celery Workers
In separate terminals, start Celery workers to execute tasks:
$ celery -A your_project worker --loglevel=info
See examples below:
Use the @block_task decorator with lambda conditions to create block processing tasks:
from abstract_block_dumper.decorators import block_task
# Process every block
@block_task(condition=lambda bn: True)
def process_every_block(block_number: int):
print(f"Processing every block: {block_number}")
# Process every 10 blocks
@block_task(condition=lambda bn: bn % 10 == 0)
def process_every_10_blocks(block_number: int):
print(f"Processing every 10 blocks: {block_number}")
# Process with multiple netuids
@block_task(
condition=lambda bn, netuid: bn % 100 == 0,
args=[{"netuid": 1}, {"netuid": 3}, {"netuid": 22}],
backfilling_lookback=300,
celery_kwargs={"queue": "high-priority"}
)
def process_multi_netuid_task(block_number: int, netuid: int):
print(f"Processing block {block_number} for netuid: {netuid}")
Maintenance Tasks
Cleanup Old Task Attempts
The framework provides a maintenance task to clean up old task records and maintain database performance:
from abstract_block_dumper.tasks import cleanup_old_tasks
# Delete tasks older than 7 days (default)
cleanup_old_tasks.delay()
# Delete tasks older than 30 days
cleanup_old_tasks.delay(days=30)
This task deletes all succeeded or unrecoverable failed tasks older than the specified number of days. It never deletes tasks with PENDING or RUNNING status to ensure ongoing work is preserved.
Running the Cleanup Task
Option 1: Manual Execution
# Using Django shell
python manage.py shell -c "from abstract_block_dumper.tasks import cleanup_old_tasks; cleanup_old_tasks.delay()"
Option 2: Cron Job (Recommended - once per day)
# Add to crontab (daily at 2 AM)
0 2 * * * cd /path/to/your/project && python manage.py shell -c "from abstract_block_dumper.tasks import cleanup_old_tasks; cleanup_old_tasks.delay()"
Option 3: Celery Beat (Automated Scheduling)
Add this to your Django settings.py:
from celery.schedules import crontab
CELERY_BEAT_SCHEDULE = {
'cleanup-old-tasks': {
'task': 'abstract_block_dumper.cleanup_old_tasks',
'schedule': crontab(hour=2, minute=0), # Daily at 2 AM
'kwargs': {'days': 7}, # Customize retention period
},
}
Then start the Celery beat scheduler:
celery -A your_project beat --loglevel=info
Configuration
Required Django Settings
Add these settings to your Django settings.py:
# Celery Configuration
CELERY_BROKER_URL = 'redis://localhost:6379/0'
CELERY_RESULT_BACKEND = 'redis://localhost:6379/0'
# Abstract Block Dumper specific settings
BITTENSOR_NETWORK = 'finney' # Options: 'finney', 'local', 'testnet', 'mainnet'
BLOCK_DUMPER_START_FROM_BLOCK = 'current' # Options: None, 'current', or int
BLOCK_DUMPER_POLL_INTERVAL = 1 # seconds between polling for new blocks
BLOCK_TASK_RETRY_BACKOFF = 2 # minutes for retry backoff base
BLOCK_DUMPER_MAX_ATTEMPTS = 3 # maximum retry attempts
BLOCK_TASK_MAX_RETRY_DELAY_MINUTES = 1440 # maximum retry delay (24 hours)
Configuration Options Reference
Core Settings
BITTENSOR_NETWORK (str, default: 'finney') Specifies which Bittensor network to connect to
BLOCK_DUMPER_START_FROM_BLOCK (str|int|None, default: None)
- Purpose: Determines the starting block for processing when the scheduler first runs
- Valid Values:
None: Resume from the last processed block stored in database'current': Start from the current blockchain block (skips historical blocks)int: Start from a specific block number (e.g.,1000000)
- Example:
BLOCK_DUMPER_START_FROM_BLOCK = 'current' - Performance Impact: Starting from historical blocks may require significant processing time
Scheduler Settings
BLOCK_DUMPER_POLL_INTERVAL (int, default: 1)
- Purpose: Seconds to wait between checking for new blocks
- Valid Range:
1to3600(1 second to 1 hour) - Example:
BLOCK_DUMPER_POLL_INTERVAL = 5 - Performance Impact:
- Lower values (1-2s): Near real-time processing, higher CPU/network usage
- Higher values (10-60s): Reduced load but delayed processing
- Very low values (<1s) may cause rate limiting
Retry and Error Handling Settings
BLOCK_DUMPER_MAX_ATTEMPTS (int, default: 3)
- Purpose: Maximum number of attempts to retry a failed task before giving up
- Valid Range:
1to10 - Example:
BLOCK_DUMPER_MAX_ATTEMPTS = 5 - Performance Impact: Higher values increase resilience but may delay failure detection
BLOCK_TASK_RETRY_BACKOFF (int, default: 1)
- Purpose: Base number of minutes for exponential backoff retry delays
- Valid Range:
1to60 - Example:
BLOCK_TASK_RETRY_BACKOFF = 2 - Calculation: Actual delay =
backoff ** attempt_countminutes- Attempt 1: 2¹ = 2 minutes
- Attempt 2: 2² = 4 minutes
- Attempt 3: 2³ = 8 minutes
- Performance Impact: Lower values retry faster but may overwhelm failing services
BLOCK_TASK_MAX_RETRY_DELAY_MINUTES (int, default: 1440)
- Purpose: Maximum delay (in minutes) between retry attempts, caps exponential backoff
- Valid Range:
1to10080(1 minute to 1 week) - Example:
BLOCK_TASK_MAX_RETRY_DELAY_MINUTES = 720# 12 hours max - Performance Impact: Prevents extremely long delays while maintaining backoff benefits
Example Project
The repository includes a complete working example in the example_project/ directory that demonstrates:
- Django application setup with abstract-block-dumper
- Multiple task types (
@every_block,@every_n_blockswith different configurations) - Error handling with a randomly failing task
- Docker Compose setup with all required services
- Monitoring with Flower (Celery monitoring tool)
Running the Example
cd example_project
docker-compose up --build
This starts:
- Django application (http://localhost:8000) - Admin interface (user:
admin, password:admin) - Celery workers - Execute block processing tasks
- Block scheduler - Monitors blockchain and schedules tasks
- Flower monitoring (http://localhost:5555) - Monitor Celery tasks
- Redis & PostgreSQL - Required services
Development
Pre-requisites:
Ideally, you should run nox -t format lint before every commit to ensure that the code is properly formatted and linted.
Before submitting a PR, make sure that tests pass as well, you can do so using:
nox -t check # equivalent to `nox -t format lint test`
If you wish to install dependencies into .venv so your IDE can pick them up, you can do so using:
uv sync --all-extras --dev
Release process
Run nox -s make_release -- X.Y.Z where X.Y.Z is the version you're releasing and follow the printed instructions.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file abstract_block_dumper-0.0.1.tar.gz.
File metadata
- Download URL: abstract_block_dumper-0.0.1.tar.gz
- Upload date:
- Size: 228.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
578ce36cb3d5857a37f1f21b486b4f99e7234d844e6da3f1488007980a8f855f
|
|
| MD5 |
eb7990a203a9b470572f17b3c4b07508
|
|
| BLAKE2b-256 |
cb79e2100ba6b8afba61e96ac88027e2f5883959cf97619fa74deb34ee7262e8
|
Provenance
The following attestation bundles were made for abstract_block_dumper-0.0.1.tar.gz:
Publisher:
publish.yml on bactensor/abstract-block-dumper
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
abstract_block_dumper-0.0.1.tar.gz -
Subject digest:
578ce36cb3d5857a37f1f21b486b4f99e7234d844e6da3f1488007980a8f855f - Sigstore transparency entry: 628535161
- Sigstore integration time:
-
Permalink:
bactensor/abstract-block-dumper@a15b7e96fa8cee24db9334a710eb13a842214da1 -
Branch / Tag:
refs/tags/v0.0.1 - Owner: https://github.com/bactensor
-
Access:
private
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@a15b7e96fa8cee24db9334a710eb13a842214da1 -
Trigger Event:
push
-
Statement type:
File details
Details for the file abstract_block_dumper-0.0.1-py3-none-any.whl.
File metadata
- Download URL: abstract_block_dumper-0.0.1-py3-none-any.whl
- Upload date:
- Size: 22.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
2f0d6afe5d2c51127c0d16b353972ea3c99e364f326f043b6f03b15ceb061fa7
|
|
| MD5 |
c149edb5f2a6c1f6eaf988bdc137087e
|
|
| BLAKE2b-256 |
ce9c2ce216d2746b9bc8d22dc3adae4036f4a5c056efefbd7525661163a1ac87
|
Provenance
The following attestation bundles were made for abstract_block_dumper-0.0.1-py3-none-any.whl:
Publisher:
publish.yml on bactensor/abstract-block-dumper
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
abstract_block_dumper-0.0.1-py3-none-any.whl -
Subject digest:
2f0d6afe5d2c51127c0d16b353972ea3c99e364f326f043b6f03b15ceb061fa7 - Sigstore transparency entry: 628535183
- Sigstore integration time:
-
Permalink:
bactensor/abstract-block-dumper@a15b7e96fa8cee24db9334a710eb13a842214da1 -
Branch / Tag:
refs/tags/v0.0.1 - Owner: https://github.com/bactensor
-
Access:
private
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@a15b7e96fa8cee24db9334a710eb13a842214da1 -
Trigger Event:
push
-
Statement type: