Skip to main content

FastAPI integration for taskiq

Project description

Taskiq + FastAPI

This repository has a code to integrate FastAPI with taskiq easily.

Taskiq and FastAPI both have dependencies and this library makes it possible to depend on fastapi.Request or starlette.requests.HTTPConnection in taskiq tasks.

With this library you can easily re-use your fastapi dependencies in taskiq functions.

How does it work?

It adds startup functions to broker so it imports your fastapi application and creates a single worker-wide Request and HTTPConnection objects that you depend on.

THIS REQUEST IS NOT RELATED TO THE ACTUAL REQUESTS IN FASTAPI! This request won't have actual data about the request you were handling while sending task.

Usage

Here we have an example of function that is being used by both taskiq's task and fastapi's handler function.

I have a script called test_script.py so my app can be found at test_script:app. We use strings to resolve application to bypass circular imports.

Also, as you can see, we use TaskiqDepends for Request. That's because taskiq dependency resolver must know that this type must be injected. FastAPI disallow Depends for Request type. That's why we use TaskiqDepends.

from fastapi import FastAPI, Request
from pydantic import BaseModel
from redis.asyncio import ConnectionPool, Redis
from fastapi import Depends as FastAPIDepends
from taskiq import TaskiqDepends
import taskiq_fastapi
from taskiq import ZeroMQBroker

broker = ZeroMQBroker()

app = FastAPI()


@app.on_event("startup")
async def app_startup():
    #####################
    # IMPORTANT NOTE    #
    #####################
    # If you won't check that this is not
    # a worker process, you'll
    # create an infinite recursion. Because in worker processes
    # fastapi startup will be called.
    if not broker.is_worker_process:
        print("Starting broker")
        await broker.startup()
    print("Creating redis pool")
    app.state.redis_pool = ConnectionPool.from_url("redis://localhost")


@app.on_event("shutdown")
async def app_shutdown():
    #####################
    # IMPORTANT NOTE    #
    #####################
    # If you won't check that this is not
    # a worker process, you'll
    # create an infinite recursion. Because in worker processes
    # fastapi startup will be called.
    if not broker.is_worker_process:
        print("Shutting down broker")
        await broker.shutdown()
    print("Stopping redis pool")
    await app.state.redis_pool.disconnect()


# Here we call our magic function.
taskiq_fastapi.init(broker, "test_script:app")


# We use TaskiqDepends here, because if we use FastAPIDepends fastapi
# initialization will fail.
def get_redis_pool(request: Request = TaskiqDepends()) -> ConnectionPool:
    return request.app.state.redis_pool


@broker.task
async def my_redis_task(
    key: str,
    val: str,
    # Here we depend using TaskiqDepends.
    # Please use TaskiqDepends for all tasks to be resolved correctly.
    # Or dependencies won't be injected.
    pool: ConnectionPool = TaskiqDepends(get_redis_pool),
):
    async with Redis(connection_pool=pool) as redis:
        await redis.set(key, val)
        print("Value set.")


class MyVal(BaseModel):
    key: str
    val: str


@app.post("/val")
async def setval_endpoint(val: MyVal) -> None:
    await my_redis_task.kiq(
        key=val.key,
        val=val.val,
    )
    print("Task sent")


@app.get("/val")
async def getval_endpoint(
    key: str,
    pool: ConnectionPool = FastAPIDepends(get_redis_pool),
) -> str:
    async with Redis(connection_pool=pool, decode_responses=True) as redis:
        return await redis.get(key)

Manually update dependency context

When using InMemoryBroker it may be required to update the dependency context manually. This may also be useful when setting up tests.

import taskiq_fastapi
from taskiq import InMemoryBroker

broker = InMemoryBroker()

app = FastAPI()

taskiq_fastapi.init(broker, "test_script:app")
taskiq_fastapi.populate_dependency_context(broker, app)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

taskiq_fastapi-0.3.2.tar.gz (5.1 kB view details)

Uploaded Source

Built Distribution

taskiq_fastapi-0.3.2-py3-none-any.whl (4.8 kB view details)

Uploaded Python 3

File details

Details for the file taskiq_fastapi-0.3.2.tar.gz.

File metadata

  • Download URL: taskiq_fastapi-0.3.2.tar.gz
  • Upload date:
  • Size: 5.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.3 CPython/3.10.12 Linux/6.5.0-1025-azure

File hashes

Hashes for taskiq_fastapi-0.3.2.tar.gz
Algorithm Hash digest
SHA256 f6f24ce07d7b784211f0d7f564d18a3d2411a5d7fa4fbb154adac449c9496e49
MD5 9097a1d8861a4ae8e255a21adcfb62b3
BLAKE2b-256 53eddc410692037802acb8f50c79811680ac3a542bdbae2d4905899d43881b38

See more details on using hashes here.

File details

Details for the file taskiq_fastapi-0.3.2-py3-none-any.whl.

File metadata

  • Download URL: taskiq_fastapi-0.3.2-py3-none-any.whl
  • Upload date:
  • Size: 4.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.3 CPython/3.10.12 Linux/6.5.0-1025-azure

File hashes

Hashes for taskiq_fastapi-0.3.2-py3-none-any.whl
Algorithm Hash digest
SHA256 85da394239801ca3b1142bf3d15ebc0a19e9e8f224f074c945b25a1c69f4e365
MD5 f73f94740ad8c5c05e140478670597f8
BLAKE2b-256 1f14c6633d983542f5d6a2774a403597292966b6a9655a34848b55072fb00cbd

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page