This is a pre-production deployment of Warehouse, however changes made here WILL affect the production instance of PyPI.
Latest Version Dependencies status unknown Test status unknown Test coverage unknown
Project Description

psq is an example Python implementation of a simple distributed task queue using Google Cloud Pub/Sub.

psq requires minimal configuration and relies on Cloud Pub/Sub to provide scalable and reliable messaging.

psq is intentionally similar to rq and simpleq, and takes some inspiration from celery and this blog post.

Installation

Install via pip:

pip install psq

Prerequisites

  • A project on the Google Developers Console.
  • The Google Cloud SDK installed locally.
  • You will need the Cloud Pub/Sub API enabled on your project. The link will walk you through enabling the API.
  • You will need to run gcloud auth before running these examples so that authentication to Google Cloud Platform services is handled transparently.

Usage

First, create a task:

def adder(a, b):
    return a + b

Then, create a pubsub client and a queue:

from google.cloud import pubsub
import psq


PROJECT_ID = 'your-project-id'

client = pubsub.Client(project=PROJECT_ID)

q = psq.Queue(client)

Now you can enqueue tasks:

from tasks import adder

q.enqueue(adder)

In order to get task results, you have to configure storage:

from google.cloud import pubsub
import psq


PROJECT_ID = 'your-project-id'

ps_client = pubsub.Client(project=PROJECT_ID)
ds_client = datastore.Client(project=PROJECT_ID)

q = psq.Queue(
    ps_client,
    storage=psq.DatastoreStorage(ds_client))

With storage configured, you can get the result of a task:

r = q.enqueue(adder, 5, 6)
r.result() # -> 11

You can also define multiple queues:

fast = psq.Queue(client, 'fast')
slow = psq.Queue(client, 'slow')

Things to note

Because psq is largely similar to rq, similar rules around tasks apply. You can put any Python function call on a queue, provided:

  • The function is importable by the worker. This means the __module__ that the function lives in must be importable. Notably, you can’t enqueue functions that are declared in the main module - such as tasks defined in a file that is run directly with python or via the interactive interpreter.
  • The function can be a string, but it must be the absolutely importable path to a function that the worker can import. Otherwise, the task will fail.
  • The worker and the applications queuing tasks must share exactly the same source code.
  • The function can’t depend on global context such as global variables, current_request, etc. Pass any needed context into the worker at queue time.

Delivery guarantees

Pub/sub guarantees your tasks will be delivered to the workers, but psq doesn’t presently guarantee that a task completes execution or exactly-once semantics, though it does allow you to provide your own mechanisms for this. This is similar to Celery’s default configuration.

Task completion guarantees can be provided via late ack support. Late ack is possible with Cloud Pub/sub, but it currently not implemented in this library. See CONTRIBUTING.md.

There are many approaches for exactly-once semantics, such as distributed locks. This is possible in systems such as zookeeper and redis.

Running a worker

Execute psqworker in the same directory where you tasks are defined:

psqworker.py config.q

psqworker only operates on one queue at a time. If you want a server to listen to multiple queues, use something like supervisord to run multiple psqworker processes.

Broadcast queues

A normal queue will send a single task to a single worker, spreading your tasks over all workers listening to the same queue. There are also broadcast queues, which will deliver a copy of the task to every worker. This is useful in situations where you want every worker to execute the same task, such as installing or upgrading software on every server.

broadcast_q = psq.BroadcastQueue(client)

def restart_apache_task():
    call(["apachectl", "restart"])

broadcast_q.enqueue(restart_apache_task)

Broadcast queues provide an implementation of the solution described in Reliable Task Scheduling on Google Compute Engine.

Note: broadcast queues do not currently support any form of storage and do not support return values.

Retries

Raising psq.Retry in your task will cause it to be retried.

from psq import Retry

def retry_if_fail(self):
    try:
        r = requests.get('http://some.flaky.service.com')
    except Exception as e:
        logging.error(e)
        raise Retry()

Flask & other contexts

You can bind an extra context manager to the queue.

app = Flask(__name__)

q = psq.Queue(extra_context=app.app_context)

This will ensure that the context is available in your tasks, which is useful for things such as database connections, etc.:

from flask import current_app

def flasky_task():
    backend = current_app.config['BACKEND']

Ideas for improvements

  • some sort of storage solution for broadcast queues.
  • Memcache/redis value store.
  • @task decorator that adds a delay/defer function.
  • Task chaining / groups / chords.
  • Late ack.
  • Gevent worker.
  • batch support for queueing.

Contributing changes

Licensing

Release History

Release History

0.5.0

This version

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

0.4.0

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

0.3.0

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

0.2.1

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

0.2.0

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

0.1.3

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

0.1.2

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

0.1.1

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

0.1.0

History Node

TODO: Figure out how to actually get changelog content.

Changelog content for this version goes here.

Donec et mollis dolor. Praesent et diam eget libero egestas mattis sit amet vitae augue. Nam tincidunt congue enim, ut porta lorem lacinia consectetur. Donec ut libero sed arcu vehicula ultricies a non tortor. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Show More

Download Files

Download Files

TODO: Brief introduction on what you do with files - including link to relevant help section.

File Name & Checksum SHA256 Checksum Help Version File Type Upload Date
psq-0.5.0-py2.py3-none-any.whl (29.4 kB) Copy SHA256 Checksum SHA256 py2.py3 Wheel Sep 20, 2016
psq-0.5.0.tar.gz (15.4 kB) Copy SHA256 Checksum SHA256 Source Sep 20, 2016

Supported By

WebFaction WebFaction Technical Writing Elastic Elastic Search Pingdom Pingdom Monitoring Dyn Dyn DNS HPE HPE Development Sentry Sentry Error Logging CloudAMQP CloudAMQP RabbitMQ Heroku Heroku PaaS Kabu Creative Kabu Creative UX & Design Fastly Fastly CDN DigiCert DigiCert EV Certificate Rackspace Rackspace Cloud Servers DreamHost DreamHost Log Hosting