Multithreaded Google Task Queue client.
Project description
# python-task-queue
Python TaskQueue object that can rapidly populate and download from cloud queues. Supports local multi-process execution as well.
# Installation
```bash
pip install numpy # make sure you do this first on a seperate line
pip install taskqueue
```
The task queue uses your CloudVolume secrets located in `$HOME/.cloudvolume/secrets/`. When using AWS SQS as your queue backend, you must provide `$HOME/.cloudvolume/secrets/aws-secret.json`. See the [CloudVolume](https://github.com/seung-lab/cloud-volume) repo for additional instructions.
The additional pip install line is to make it easier for CloudVolume to install as this library uses its facilities for accessing secrets.
# Usage
Define a class that inherits from taskqueue.RegisteredTask and implments the `execute` method.
Tasks can be loaded into queues locally or as based64 encoded data in the cloud and executed later.
Here's an example implementation of a `PrintTask`. Generally, you should specify a very lightweight
container and let the actual execution download and manipulate data.
```python
from taskqueue import RegisteredTask
class PrintTask(RegisteredTask):
def __init__(self, txt=''):
super(PrintTask, self).__init__(txt)
self.txt = txt
def execute(self):
if self.txt:
print(str(self) + ": " + str(self.txt))
else:
print(self)
```
## Local Usage
For small jobs, you might want to use one or more processes to execute the tasks:
```python
from taskqueue import LocalTaskQueue
with LocalTaskQueue(parallel=5) as tq: # use 5 processes
for _ in range(1000):
tq.insert(
PrintTask(i)
)
```
This will load the queue with 1000 print tasks then execute them across five processes.
## Cloud Usage
Set up an SQS queue and acquire an aws-secret.json that is compatible with CloudVolume.
```python
from taskqueue import TaskQueue
qurl = 'https://sqs.us-east-1.amazonaws.com/$DIGITS/$QUEUE_NAME'
with TaskQueue(queue_server='sqs', qurl=qurl) as tq:
for _ in range(1000):
tq.insert(PrintTask(i))
```
This inserts 1000 PrintTask descriptions into your SQS queue.
Somewhere else, you'll do the following (probably across multiple workers):
```python
from taskqueue import TaskQueue
qurl = 'https://sqs.us-east-1.amazonaws.com/$DIGITS/$QUEUE_NAME'
with TaskQueue(queue_server='sqs', qurl=qurl) as tq:
task = tq.lease(seconds=int($LEASE_SECONDS))
task.execute()
tq.delete(task)
```
Python TaskQueue object that can rapidly populate and download from cloud queues. Supports local multi-process execution as well.
# Installation
```bash
pip install numpy # make sure you do this first on a seperate line
pip install taskqueue
```
The task queue uses your CloudVolume secrets located in `$HOME/.cloudvolume/secrets/`. When using AWS SQS as your queue backend, you must provide `$HOME/.cloudvolume/secrets/aws-secret.json`. See the [CloudVolume](https://github.com/seung-lab/cloud-volume) repo for additional instructions.
The additional pip install line is to make it easier for CloudVolume to install as this library uses its facilities for accessing secrets.
# Usage
Define a class that inherits from taskqueue.RegisteredTask and implments the `execute` method.
Tasks can be loaded into queues locally or as based64 encoded data in the cloud and executed later.
Here's an example implementation of a `PrintTask`. Generally, you should specify a very lightweight
container and let the actual execution download and manipulate data.
```python
from taskqueue import RegisteredTask
class PrintTask(RegisteredTask):
def __init__(self, txt=''):
super(PrintTask, self).__init__(txt)
self.txt = txt
def execute(self):
if self.txt:
print(str(self) + ": " + str(self.txt))
else:
print(self)
```
## Local Usage
For small jobs, you might want to use one or more processes to execute the tasks:
```python
from taskqueue import LocalTaskQueue
with LocalTaskQueue(parallel=5) as tq: # use 5 processes
for _ in range(1000):
tq.insert(
PrintTask(i)
)
```
This will load the queue with 1000 print tasks then execute them across five processes.
## Cloud Usage
Set up an SQS queue and acquire an aws-secret.json that is compatible with CloudVolume.
```python
from taskqueue import TaskQueue
qurl = 'https://sqs.us-east-1.amazonaws.com/$DIGITS/$QUEUE_NAME'
with TaskQueue(queue_server='sqs', qurl=qurl) as tq:
for _ in range(1000):
tq.insert(PrintTask(i))
```
This inserts 1000 PrintTask descriptions into your SQS queue.
Somewhere else, you'll do the following (probably across multiple workers):
```python
from taskqueue import TaskQueue
qurl = 'https://sqs.us-east-1.amazonaws.com/$DIGITS/$QUEUE_NAME'
with TaskQueue(queue_server='sqs', qurl=qurl) as tq:
task = tq.lease(seconds=int($LEASE_SECONDS))
task.execute()
tq.delete(task)
```
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
task-queue-0.7.0.tar.gz
(10.5 kB
view details)
File details
Details for the file task-queue-0.7.0.tar.gz
.
File metadata
- Download URL: task-queue-0.7.0.tar.gz
- Upload date:
- Size: 10.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.2.0 requests-toolbelt/0.8.0 tqdm/4.23.0 CPython/3.6.3
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | d67e9718c50e1cfa4ceef9e7619ac5a985c0751aadb166a418b80edc92c9c7b2 |
|
MD5 | e9c371e18d594cba95d27de8f927d38b |
|
BLAKE2b-256 | 3b442622d81ea78f21858fba487ae4bca49907add0da483d8594458f621fd283 |