Skip to main content

Microservice library on asyncio - HTTP server, websockets, pub/sub messaging for AWS SNS+SQS and RabbitMQ

Project description

tomodachi - a lightweight microservices library on Python asyncio

A Python 3 microservice library / framework using asyncio (async / await) with HTTP, websockets, RabbitMQ / AMQP and AWS SNS+SQS built-in support for event based messaging and intra-service communication.

https://github.com/kalaspuff/tomodachi/workflows/Python%20package/badge.svg https://img.shields.io/pypi/v/tomodachi.svg https://codecov.io/gh/kalaspuff/tomodachi/branch/master/graph/badge.svg https://img.shields.io/pypi/pyversions/tomodachi.svg

Tomodachi is a tiny framework designed to build fast microservices listening on HTTP or communicating over event driven message buses like RabbitMQ, AMQP, AWS (Amazon Web Services) SNS+SQS, etc. It’s designed to be extendable to make use of any type of transport layer available.

Tomodachi [友達] means friends – a suitable name for microservices working together. 😻 👬 👭 👫 😻

Project documentation

https://img.shields.io/badge/tomodachi.dev-documentation-ff69b4

Usage

tomodachi is used to execute service code via command line interface or within container images.

Usage: tomodachi <command> [options] [arguments]

Options:
  -h, --help                                Show this help message and exit
  -v, --version                             Print tomodachi version
  --dependency-versions                     Print versions of dependencies

Available commands:
  ---
  Command: run
  Starts service(s) defined in the .py files specified as <service> argument(s)

  $ tomodachi run <service ...> [-c <config-file ...>] [--production]
  | --loop [auto|asyncio|uvloop]            Event loop implementation [asyncio]
  | --production                            Disable restart on file changes
  | -c, --config <files>                    Use configuration from JSON files
  | -l, --log <level>, --log-level <level>  Specify log level
https://raw.githubusercontent.com/kalaspuff/tomodachi/master/docs/assets/microservice-in-30-seconds-white.gif

README

This documentation README includes a guide of how to get started with services, what built-in functionality exists in this library, lists of available configuration parameters and a few examples of service code.

Use https://tomodachi.dev/docs for extensive project documentation.


Please note: this library is a work in progress.

Consider tomodachi as beta software. tomodachi is still an experimental project with an unregular release schedule. The package is not yet available as 1.0.0 and there may be breaking changes between 0.x versions.


Getting started 🏃

First off – installation using poetry is fully supported and battle-tested (pip works just as fine)

Install tomodachi in your preferred way, wether it be poetry, pip, pipenv, etc. Installing the distribution will give your environment access to the tomodachi package for imports as well as a shortcut to the CLI alias, which later is used to run the microservices you build.

local ~$ pip install tomodachi
> ...
> Installing collected packages: ..., ..., ..., tomodachi
> Successfully installed ... ... ... tomodachi-x.x.xx

local ~$ tomodachi --version
> tomodachi x.xx.xx

Probably goes without saying – services you build, their dependencies, together with runtime utilities like this one, should preferably always be installed and run in isolated environments like Docker containers or virtual environments.

Building blocks for a service class and microservice entrypoint

  1. import tomodachi and create a class that inherits tomodachi.Service, it can be called anything… or just Service to keep it simple.

  2. Add a name attribute to the class and give it a string value. Having a name attribute isn’t required, but good practice.

  3. Define an awaitable function in the service class – in this example we’ll use it as an entrypoint to trigger code in the service by decorating it with one of the available invoker decorators. Note that a service class must have at least one decorated function available to even be recognized as a service by tomodachi run.

  4. Decide on how to trigger the function – for example using HTTP, pub/sub or on a timed interval, then decorate your function with one of these trigger / subscription decorators, which also invokes what capabilities the service initially has.

Further down you’ll find a desciption of how each of the built-in invoker decorators work and which keywords and parameters you can use to change their behaviour.

Note: Publishing and subscribing to events and messages may require user credentials or hosting configuration to be able to access queues and topics.

For simplicity, let’s do HTTP:

  • On each POST request to /sheep, the service will wait for up to one whole second (pretend that it’s performing I/O – waiting for response on a slow sheep counting database modification, for example) and then issue a 200 OK with some data.

  • It’s also possible to query the amount of times the POST tasks has run by doing a GET request to the same url, /sheep.

  • By using @tomodachi.http an HTTP server backed by aiohttp will be started on service start. tomodachi will act as a middleware to route requests to the correct handlers, upgrade websocket connections and then also gracefully await connections with still executing tasks, when the service is asked to stop – up until a configurable amount of time has passed.

import asyncio
import random

import tomodachi


class Service(tomodachi.Service):
    name = "sleepy-sheep-counter"

    _sheep_count = 0

    @tomodachi.http("POST", r"/sheep")
    async def add_to_sheep_count(self, request):
        await asyncio.sleep(random.random())
        self._sheep_count += 1
        return 200, str(self._sheep_count)

    @tomodachi.http("GET", r"/sheep")
    async def return_sheep_count(self, request):
        return 200, str(self._sheep_count)

Run services with:

local ~/code/service$ tomodachi run <path to .py file with service class code>

Beside the currently existing built-in ways of interfacing with a service, it’s possible to build additional function decorators to suit the use-cases one may have.

To give a few possible examples / ideas of functionality that could be coded to call functions with data in similar ways:

  • Using Redis as a task queue with configurable keys to push or pop onto.

  • Subscribing to Kinesis or Kafka event streams and act on the data received.

  • An abstraction around otherwise complex functionality or to unify API design.

  • As an example to above sentence; GraphQL resolver functionality with built-in tracability and authentication management, with a unified API to application devs.


Additional examples will follow with different ways to trigger functions in the service.

Of course the different ways can be used within the same class, for example the very common use-case of having a service listening on HTTP while also performing some kind of async pub/sub tasks.

Basic HTTP based service 🌟

Code for a simple service which would service data over HTTP, pretty similar, but with a few more concepts added.

import tomodachi


class Service(tomodachi.Service):
    name = "http-example"

    # Request paths are specified as regex for full flexibility
    @tomodachi.http("GET", r"/resource/(?P<id>[^/]+?)/?")
    async def resource(self, request, id):
        # Returning a string value normally means 200 OK
        return f"id = {id}"

    @tomodachi.http("GET", r"/health")
    async def health_check(self, request):
        # Return can also be a tuple, dict or even an aiohttp.web.Response
        # object for more complex responses - for example if you need to
        # send byte data, set your own status code or define own headers
        return {
            "body": "Healthy",
            "status": 200,
        }

    # Specify custom 404 catch-all response
    @tomodachi.http_error(status_code=404)
    async def error_404(self, request):
        return "error 404"

RabbitMQ or AWS SNS+SQS event based messaging service 🐰

Example of a service that calls a function when messages are published on an AMQP topic exchange.

import tomodachi


class Service(tomodachi.Service):
    name = "amqp-example"

    # The "message_envelope" attribute can be set on the service class to build / parse data.
    # message_envelope = ...

    # A route / topic on which the service will subscribe to via RabbitMQ / AMQP
    @tomodachi.amqp("example.topic")
    async def example_func(self, message):
        # Received message, fordarding the same message as response on another route / topic
        await tomodachi.amqp_publish(self, message, routing_key="example.response")

AWS SNS+SQS event based messaging service 📡

Example of a service using AWS SNS+SQS managed pub/sub messaging. AWS SNS and AWS SQS together brings managed message queues for microservices, distributed systems, and serverless applications hosted on AWS. tomodachi services can customize their enveloping functionality to both unwrap incoming messages and/or to produce enveloped messages for published events / messages. Pub/sub patterns are great for scalability in distributed architectures, when for example hosted in Docker on Kubernetes.

import tomodachi


class Service(tomodachi.Service):
    name = "aws-example"

    # The "message_envelope" attribute can be set on the service class to build / parse data.
    # message_envelope = ...

    # Using the @tomodachi.aws_sns_sqs decorator to make the service create an AWS SNS topic,
    # an AWS SQS queue and to make a subscription from the topic to the queue as well as start
    # receive messages from the queue using SQS.ReceiveMessages.
    @tomodachi.aws_sns_sqs("example-topic", queue_name="example-queue")
    async def example_func(self, message):
        # Received message, forwarding the same message as response on another topic
        await tomodachi.aws_sns_sqs_publish(self, message, topic="another-example-topic")

Scheduling, inter-communication between services, etc. ⚡️

There are other examples available with code of how to use services with self-invoking methods called on a specified interval or at specific times / days, as well as additional examples for inter-communication pub/sub between different services on both AMQP or AWS SNS+SQS as shown above. See more at the examples folder.


Run the service 😎

# cli alias is set up automatically on installation
local ~/code/service$ tomodachi run service.py

# shortcut to cli endpoint could be used if cloned from repo and not yet installed
local ~/code/tomodachi$ python tomodachi.py run example/http_simple_service.py

Defaults to output information on stdout.

local ~/code/service$ tomodachi run service.py
>
> ---
> Starting tomodachi services (pid: 1) ...
> * service.py
>
> Current version: tomodachi x.x.xx on Python 3.x.x
> Event loop implementation: asyncio
> Local time: October 04, 2020 - 13:38:01,201509 UTC
> Timestamp in UTC: 2020-10-04T13:38:01.201509Z
>
> File watcher is active - code changes will automatically restart services
> Quit running services with <ctrl+c>
>
> 2020-10-04 13:38:01,234 (services.service): Initializing service "example" [id: <uuid>]
> 2020-10-04 13:38:01,248 (transport.http): Listening [http] on http://127.0.0.1:9700/
> 2020-10-04 13:38:01,248 (services.service): Started service "example" [id: <uuid>]

HTTP service acts like a normal web server.

local ~$ curl -v "http://127.0.0.1:9700/resource/1234"
> HTTP/1.1 200 OK
> Content-Type: text/plain; charset=utf-8
> Server: tomodachi
> Content-Length: 9
> Date: Mon, 02 Oct 2017 13:38:02 GMT
>
> id = 1234

Example of a microservice containerized in Docker 🐳

A great way to distribute and operate microservices are usually to run them in containers or even more interestingly, in clusters of compute nodes. Here follows an example of getting a tomodachi based service up and running in Docker.

We’re building the service’ container image using just two small files, the Dockerfile and the actual code for the microservice, service.py. In reality a service would probably not be quite this small, but as a template to get started.

Dockerfile

FROM python:3.8-slim
RUN pip install tomodachi
RUN mkdir /app
WORKDIR /app
COPY service.py .
ENV PYTHONUNBUFFERED=1
CMD ["tomodachi", "run", "service.py", "--production"]

service.py

import json

import tomodachi


class Service(tomodachi.Service):
    name = "example"
    options = {
        "http": {
            "port": 80,
            "content_type": "application/json; charset=utf-8"
        }
    }

    _healthy = True

    @tomodachi.http("GET", r"/")
    async def index_endpoint(self, request):
        # tomodachi.get_execution_context() can be used for
        # debugging purposes or to add additional service context
        # in logs or alerts.
        execution_context = tomodachi.get_execution_context()

        return json.dumps({
            "data": "hello world!",
            "execution_context": execution_context,
        })

    @tomodachi.http("GET", r"/health/?", ignore_logging=True)
    async def health_check(self, request):
        if self._healthy:
            return 200, json.dumps({"status": "healthy"})
        else:
            return 503, json.dumps({"status": "not healthy"})

    @tomodachi.http_error(status_code=400)
    async def error_400(self, request):
        return json.dumps({"error": "bad-request"})

    @tomodachi.http_error(status_code=404)
    async def error_404(self, request):
        return json.dumps({"error": "not-found"})

    @tomodachi.http_error(status_code=405)
    async def error_405(self, request):
        return json.dumps({"error": "method-not-allowed"})

Building and running the container, forwarding host’s port 31337 to port 80.

local ~/code/service$ docker build . -t tomodachi-microservice
> Sending build context to Docker daemon  9.216kB
> Step 1/7 : FROM python:3.10-slim
> 3.8-slim: Pulling from library/python
> ...
>  ---> 3f7f3ab065d4
> Step 7/7 : CMD ["tomodachi", "run", "service.py", "--production"]
>  ---> Running in b8dfa9deb243
> Removing intermediate container b8dfa9deb243
>  ---> 8f09a3614da3
> Successfully built 8f09a3614da3
> Successfully tagged tomodachi-microservice:latest
local ~/code/service$ docker run -ti -p 31337:80 tomodachi-microservice
> 2020-10-04 13:38:01,234 (services.service): Initializing service "example" [id: <uuid>]
> 2020-10-04 13:38:01,248 (transport.http): Listening [http] on http://127.0.0.1:80/
> 2020-10-04 13:38:01,248 (services.service): Started service "example" [id: <uuid>]

Making requests to the running container.

local ~$ curl http://127.0.0.1:31337/ | jq
> {
>   "data": "hello world!",
>   "execution_context": {
>     "tomodachi_version": "x.x.xx",
>     "python_version": "3.x.x",
>     "system_platform": "Linux",
>     "process_id": 1,
>     "init_timestamp": "2020-10-04T13:38:01.201509Z",
>     "event_loop": "asyncio",
>     "http_enabled": true,
>     "http_current_tasks": 1,
>     "http_total_tasks": 1,
>     "aiohttp_version": "x.x.xx"
>   }
> }

local ~$ curl http://127.0.0.1:31337/health -i
> HTTP/1.1 200 OK
> Content-Type: application/json; charset=utf-8
> Server: tomodachi
> Content-Length: 21
> Date: Sun, 04 Oct 2020 13:40:44 GMT
>
> {"status": "healthy"}

local ~$ curl http://127.0.0.1:31337/no-route -i
> HTTP/1.1 404 Not Found
> Content-Type: application/json; charset=utf-8
> Server: tomodachi
> Content-Length: 22
> Date: Sun, 04 Oct 2020 13:41:18 GMT
>
> {"error": "not-found"}

It’s actually as easy as that to get something spinning. The hard part is usually to figure out (or decide) what to build next.

Other popular ways of running microservices are of course to use them as serverless functions, with an ability of scaling to zero (Lambda, Cloud Functions, Knative, etc. may come to mind). Currently tomodachi works best in a container setup and until proper serverless supporting execution context is available in the library, it should be adviced to hold off and use other tech for those kinds of deployments.


Available built-ins used as endpoints 🚀

As shown, there’s different ways to trigger your microservice function in which the most common ones are either directly via HTTP or via event based messaging (for example AMQP or AWS SNS+SQS). Here’s a list of the currently available built-ins you may use to decorate your service functions.

HTTP endpoints:

@tomodachi.http(method, url, ignore_logging=[200])

Sets up an HTTP endpoint for the specified method (GET, PUT, POST, DELETE) on the regexp url. Optionally specify ignore_logging as a dict or tuple containing the status codes you do not wish to log the access of. Can also be set to True to ignore everything except status code 500.

@tomodachi.http_static(path, url)

Sets up an HTTP endpoint for static content available as GET / HEAD from the path on disk on the base regexp url.

@tomodachi.websocket(url)

Sets up a websocket endpoint on the regexp url. The invoked function is called upon websocket connection and should return a two value tuple containing callables for a function receiving frames (first callable) and a function called on websocket close (second callable). The passed arguments to the function beside the class object is first the websocket response connection which can be used to send frames to the client, and optionally also the request object.

@tomodachi.http_error(status_code)

A function which will be called if the HTTP request would result in a 4XX status_code. You may use this for example to set up a custom handler on “404 Not Found” or “403 Forbidden” responses.

AWS SNS+SQS messaging:

@tomodachi.aws_sns_sqs(topic, competing=True, queue_name=None, filter_policy=None, **kwargs)

This would set up an AWS SQS queue, subscribing to messages on the AWS SNS topic topic, whereafter it will start consuming messages from the queue.

The competing value is used when the same queue name should be used for several services of the same type and thus “compete” for who should consume the message. Since tomodachi version 0.19.x this value has a changed default value and will now default to True as this is the most likely use-case for pub/sub in distributed architectures.

Unless queue_name is specified an auto generated queue name will be used. Additional prefixes to both topic and queue_name can be assigned by setting the options.aws_sns_sqs.topic_prefix and options.aws_sns_sqs.queue_name_prefix dict values.

The filter_policy value of specified as a keyword argument will be applied on the SNS subscription (for the specified topic and queue) as the "FilterPolicy attribute. This will apply a filter on SNS messages using the chosen “message attributes” and/or their values specified in the filter. Make note that the filter policy dict structure differs somewhat from the actual message attributes, as values to the keys in the filter policy must be a dict (object) or list (array). Example: A filter policy value of {"event": ["order_paid"], "currency": ["EUR", "USD"]} would set up the SNS subscription to receive messages on the topic only where the message attribute "event" is "order_paid" and the "currency" value is either "EUR" or "USD".

If filter_policy is not specified as an argument (default), the queue will receive messages on the topic as per already specified if using an existing subscription, or receive all messages on the topic if a new subscription is set up (default). Changing the filter_policy on an existing subscription may take several minutes to propagate. Read more about the filter policy format on AWS. https://docs.aws.amazon.com/sns/latest/dg/sns-subscription-filter-policies.html

Related to the above mentioned filter policy, the aws_sns_sqs_publish function (which is used for publishing messages) can specify “message attributes” using the message_attributes keyword argument. Values should be specified as a simple dict with keys and values. Example: {"event": "order_paid", "paid_amount": 100, "currency": "EUR"}.

Depending on the service message_envelope (previously named message_protocol) attribute if used, parts of the enveloped data would be distributed to different keyword arguments of the decorated function. It’s usually safe to just use data as an argument. You can also specify a specific message_envelope value as a keyword argument to the decorator for specifying a specific enveloping method to use instead of the global one set for the service.

If you’re utilizing from tomodachi.envelope import ProtobufBase and using ProtobufBase as the specified service message_envelope you may also pass a keyword argument proto_class into the decorator, describing the protobuf (Protocol Buffers) generated Python class to use for decoding incoming messages. Custom enveloping classes can be built to fit your existing architecture or for even more control of tracing and shared metadata between services.

AMQP messaging (RabbitMQ):

@tomodachi.amqp(routing_key, exchange_name="amq.topic", competing=True, queue_name=None, **kwargs)

Sets up the method to be called whenever a AMQP / RabbitMQ message is received for the specified routing_key. By default the 'amq.topic' topic exchange would be used, it may also be overridden by setting the options.amqp.exchange_name dict value on the service class.

The competing value is used when the same queue name should be used for several services of the same type and thus “compete” for who should consume the message. Since tomodachi version 0.19.x this value has a changed default value and will now default to True as this is the most likely use-case for pub/sub in distributed architectures.

Unless queue_name is specified an auto generated queue name will be used. Additional prefixes to both routing_key and queue_name can be assigned by setting the options.amqp.routing_key_prefix and options.amqp.queue_name_prefix dict values.

Depending on the service message_envelope (previously named message_protocol) attribute if used, parts of the enveloped data would be distributed to different keyword arguments of the decorated function. It’s usually safe to just use data as an argument. You can also specify a specific message_envelope value as a keyword argument to the decorator for specifying a specific enveloping method to use instead of the global one set for the service.

If you’re utilizing from tomodachi.envelope import ProtobufBase and using ProtobufBase as the specified service message_envelope you may also pass a keyword argument proto_class into the decorator, describing the protobuf (Protocol Buffers) generated Python class to use for decoding incoming messages. Custom enveloping classes can be built to fit your existing architecture or for even more control of tracing and shared metadata between services.

Scheduled functions / cron / triggered on time interval:

@tomodachi.schedule(interval=None, timestamp=None, timezone=None, immediately=False)

A scheduled function invoked on either a specified interval (you may use the popular cron notation as a str for fine-grained interval or specify an integer value of seconds) or a specific timestamp. The timezone will default to your local time unless explicitly stated.

When using an integer interval you may also specify wether the function should be called immediately on service start or wait the full interval seconds before its first invokation.

@tomodachi.heartbeat

A function which will be invoked every second.

@tomodachi.minutely, @tomodachi.hourly, @tomodachi.daily, @tomodachi.monthly

A scheduled function which will be invoked once every minute / hour / day / month.

A word on scheduled tasks in distributed contexts: What is your use-case for scheduling function triggers or functions that trigger on an interval. These types of scheduling may not be optimal in clusters with many pods in the same replication set, as all the services running the same code will very likely execute at the same timestamp / interval (which in same cases may correlated with exactly when they were last deployed). As such these functions are quite naive and should only be used with some care, so that it triggering the functions several times doesn’t incur unnecessary costs or come as a bad surprise if the functions aren’t completely idempotent. To perform a task on a specific timestamp or on an interval where only one of the available services of the same type in a cluster should trigger is a common thing to solve and there are several solutions to pick from., some kind of distributed consensus needs to be reached. Tooling exists, but what you need may differ depending on your use-case. There’s algorithms for distributed consensus and leader election, Paxos or Raft, that luckily have already been implemented to solutions like the strongly consistent and distributed key-value stores etcd and TiKV. Even primitive solutions such as Redis SETNX commands would work, but could be costly or hard to manage access levels around. If you’re on k8s there’s even a simple “leader election” API available that just creates a 15 seconds lease. Solutions are many and if you are in need, go hunting and find one that suits your use-case, there’s probably tooling and libraries available to call it from your service functions.

Implementing proper consensus mechanisms and in turn leader election can be complicated. In distributed environments the architecture around these solutions needs to account for leases, decision making when consensus was not reached, how to handle crashed executors, quick recovery on master node(s) disruptions, etc.


To extend the functionality by building your own trigger decorators for your endpoints, studying the built-in invoker classes should the first step of action. All invoker classes should extend the class for a common developer experience: tomodachi.invoker.Invoker.


Additional configuration options 🤩

A tomodachi.Service extended service class may specify a class attribute named options (as a dict) for additional configuration.

import json

import tomodachi


class Service(tomodachi.Service):
    name = "http-example"
    options = {
        "http": {
            "port": 80,
            "content_type": "application/json; charset=utf-8",
            "real_ip_from": [
                "127.0.0.1/32",
                "10.0.0.0/8",
                "172.16.0.0/12",
                "192.168.0.0/16"
            ],
            "keepalive_timeout": 5,
            "max_keepalive_requests": 20,
        },
        "watcher": {
            "ignored_dirs": ["node_modules"],
        },
    }

    @tomodachi.http("GET", r"/health")
    async def health_check(self, request):
        return 200, json.dumps({"status": "healthy"})

    # Specify custom 404 catch-all response
    @tomodachi.http_error(status_code=404)
    async def error_404(self, request):
        return json.dumps({"error": "not-found"})

⁝⁝ HTTP server parameters ⁝⁝ options["http"][key]

_____________________________

Configuration key

Description

Default value

http.port

TCP port (integer value) to listen for incoming connections.

9700

http.host

Network interface to bind TCP server to. "0.0.0.0" will bind to all IPv4 interfaces. None or "" will assume all network interfaces.

"0.0.0.0"

http.keepalive_timeout

Enables connections to use keep-alive if set to an integer value over 0. Number of seconds to keep idle incoming connections open.

0

http.max_keepalive_requests

An optional number (int) of requests which is allowed for a keep-alive connection. After the specified number of requests has been done, the connection will be closed. An option value of 0 or None (default) will allow any number of requests over an open keep-alive connection.

None

http.max_keepalive_time

An optional maximum time in seconds (int) for which keep-alive connections are kept open. If a keep-alive connection has been kept open for more than http.max_keepalive_time seconds, the following request will be closed upon returning a response. The feature is not used by default and won’t be used if the value is 0 or None. A keep-alive connection may otherwise be open unless inactive for more than the keep-alive timeout.

None

http.client_max_size

The client’s maximum size in a request, as an integer, in bytes.

(1024 ** 2) * 100

http.termination_grace_period_seconds

The number of seconds to wait for functions called via HTTP to gracefully finish execution before terminating the service, for example if service received a SIGINT or SIGTERM signal while requests were still awaiting response results.

30

http.real_ip_header

Header to read the value of the client’s real IP address from if service operates behind a reverse proxy. Only used if http.real_ip_from is set and the proxy’s IP correlates with the value from http.real_ip_from.

"X-Forwarded-For"

http.real_ip_from

IP address(es) or IP subnet(s) / CIDR. Allows the http.real_ip_header header value to be used as client’s IP address if connecting reverse proxy’s IP equals a value in the list or is within a specified subnet. For example ["127.0.0.1/32", "10.0.0.0/8", "172.16.0.0/12", "192.168.0.0/16"] would permit header to be used if closest reverse proxy is "127.0.0.1" or within the three common private network IP address ranges.

[]

http.content_type

Default content-type header to use if not specified in the response.

"text/plain; charset=utf-8"

http.access_log

If set to the default value (boolean) True the HTTP access log will be output to stdout (logger transport.http). If set to a str value, the access log will additionally also be stored to file using value as filename.

True

http.server_header

"Server" header value in responses.

"tomodachi"

⁝⁝ Credentials and prefixes for AWS SNS+SQS pub/sub ⁝⁝ options["aws_sns_sqs"][key]

Configuration key

Description

Default value

aws_sns_sqs.region_name

None

aws_sns_sqs.aws_access_key_id

None

aws_sns_sqs.aws_secret_access_key

None

aws_sns_sqs.topic_prefix

""

aws_sns_sqs.queue_name_prefix

""

⁝⁝ Configure custom AWS endpoints for development ⁝⁝ options["aws_endpoint_urls"][key]

Configuration key

Description

Default value

aws_endpoint_urls.sns

None

aws_endpoint_urls.sqs

None

⁝⁝ AMQP / RabbitMQ pub/sub settings ⁝⁝ options["amqp"][key]

Configuration key

Description

Default value

amqp.host

"127.0.0.1"

amqp.port

5672

amqp.login

"guest"

amqp.password

"guest"

amqp.exchange_name

"amq_topic"

amqp.routing_key_prefix

""

amqp.queue_name_prefix

""

amqp.virtualhost

"/"

amqp.ssl

False

amqp.heartbeat

60

amqp.queue_ttl

86400

⁝⁝ Options for code auto reload on file changes in development ⁝⁝ options["watcher"][key]

Configuration key

Description

Default value

watcher.ignored_dirs

[]

watcher.watched_file_endings

[]

Decorated functions using @tomodachi.decorator 🎄

Invoker functions can of course be decorated using custom functionality. For ease of use you can then in turn decorate your decorator with the the built-in @tomodachi.decorator to ease development. If the decorator would return anything else than True or None (or not specifying any return statement) the invoked function will not be called and instead the returned value will be used, for example as an HTTP response.

import tomodachi


@tomodachi.decorator
async def require_csrf(instance, request):
    token = request.headers.get("X-CSRF-Token")
    if not token or token != request.cookies.get("csrftoken"):
        return {
            "body": "Invalid CSRF token",
            "status": 403
        }


class Service(tomodachi.Service):
    name = "example"

    @tomodachi.http("POST", r"/create")
    @require_csrf
    async def create_data(self, request):
        # Do magic here!
        return "OK"

Requirements 👍

  • Python (3.7+, 3.8+, 3.9+)

  • aiohttp (aiohttp is the currently supported HTTP server implementation for tomodachi)

  • aiobotocore and botocore (used for AWS SNS+SQS pub/sub messaging)

  • aioamqp (used for RabbitMQ / AMQP pub/sub messaging)

  • uvloop (optional: alternative event loop implementation)

LICENSE 🙋

tomodachi is offered under the MIT License.

CHANGELOG 🧳

Changes are recorded in the repo as well as together with the GitHub releases.

GITHUB / SOURCE 🦄

The latest developer version of tomodachi is always available at GitHub.

Any questions?

What is the best way to run a tomodachi service?

Docker containers are great and can be scaled out in Kubernetes, Nomad or other orchestration engines. Some may instead run several services on the same environment, on the same machine if their workloads are smaller or more consistent. Remember to gather your output and monitor your instances or clusters.

For real workloads: Go for a Dockerized environment if possible – async task queues are usually nice and services could scale up and down for keeping up with incoming demand; if you require network access like HTTP from users or API clients directly to the service, then it’s usually preferred to put some kind of ingress (nginx, haproxy or other type of load balancer) to proxy requests to the service pods. Let the ingress then handle public TLS, http2 / http3, client facing keep-alives and WebSocket protocol upgrades and let the service instead take care of the business logic.

Are there any more example services?

There are a few examples in the examples folder, including using tomodachi in an example Docker environment with or without docker-compose. There are examples to publish events / messages to an AWS SNS topic and subscribe to an AWS SQS queue. There’s also a similar code available of how to work with pub/sub for RabbitMQ via the AMQP transport protocol.

Why should I use this?

tomodachi is a perfect place to start when experimenting with your architecture or trying out a concept for a new service. It may not have all the features you desire and it may never do, but I believe it’s great for bootstrapping microservices in async Python.

I have some great additions!

Sweet! Please send me a PR with your ideas. There’s now automatic tests that are running as GitHub actions to verify linting and regressions. Get started at the short contribution guide.

Beta software in production?

There are some projects and organizations that already are running services based on tomodachi in production. The library is provided as is with an unregular release schedule, and as with most software, there will be unfortunate bugs or crashes. Consider this currently as beta software (with an ambition to be stable enough for production). Would be great to hear about other use-cases in the wild!

Another good idea is to drop in Sentry or other exception debugging solutions. These are great to catch errors if something wouldn’t work as expected in the internal routing or if your service code raises unhandled exceptions.

Who built this and why?

My name is Carl Oscar Aaro [@kalaspuff] and I’m a coder from Sweden. When I started writing the first few lines of this library back in 2016, my intention was just to learn more about Python’s asyncio, the event loop, event sourcing and message queues. A lot has happened since – now running services in both production and development clusters, while also using microservices for quick proof of concepts and experimentation. 🎉

Contributions

Please help out to add features that you deem are missing and/or fix bugs in the repo.

To add a PR, for the repository, commit your changes to your own clone and make a PR on GitHub for your clone against master branch.

Read more in the contribution guide.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tomodachi-0.20.3.tar.gz (97.2 kB view hashes)

Uploaded Source

Built Distribution

tomodachi-0.20.3-py3-none-any.whl (82.7 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page