Skip to main content

A comprehensive Redis toolkit for Python with caching, memoization, and utilities

Project description

rediskit

A Python toolkit that provides Redis-backed performance and concurrency primitives for applications. It enables developers to add caching, distributed coordination, and data protection to their Python applications with minimal effort.

Still work in progress

Many features are still under development, there will be many breaking changes. Please use at your own risk.

Features

  • Function Result Caching: Use the @RedisMemoize decorator to cache expensive function calls with automatic serialization, compression, and encryption
  • Distributed Coordination: Redis-based distributed locks and semaphores for coordinating access across multiple processes/machines
  • Data Protection: Multi-version encryption keys with automatic key rotation for sensitive cached data
  • Async Support: Full support for both synchronous and asynchronous applications
  • Flexible Storage: Choose between string or hash-based Redis storage patterns
  • Modern Type Hints: Full type safety with Python 3.12+ syntax

Installation

uv add rediskit
# or
poetry add rediskit

Quick Start

Basic Setup

from rediskit import redis_memoize, init_redis_connection_pool

# Initialize Redis connection pool (call once at app startup)
init_redis_connection_pool()


# Cache expensive function results
@redis_memoize(memoize_key="expensive_calc", ttl=300)
def expensive_calculation(tenantId: str, value: int) -> dict:
    # Simulate expensive computation
    import time
    time.sleep(2)
    return {"result": value * 42}


# Usage
result = expensive_calculation("tenant1", 10)  # Takes 2 seconds
result = expensive_calculation("tenant1", 10)  # Returns instantly from cache

Custom Redis Connection

import redis
from rediskit import redis_memoize

# Use your own Redis connection
my_redis = redis.Redis(host='my-redis-host', port=6379, db=1)


@redis_memoize(
    memoize_key="custom_calc",
    ttl=600,
    connection=my_redis
)
def my_function(tenantId: str, data: dict) -> dict:
    return {"processed": data}

Advanced Caching Options

from rediskit import redis_memoize


# Hash-based storage with encryption
@redis_memoize(
    memoize_key=lambda tenantId, user_id: f"user_profile:{tenantId}:{user_id}",
    ttl=3600,
    storage_type="hash",  # Store in Redis hash for efficient field access
    enable_encryption=True,  # Encrypt sensitive data
    cache_type="zipJson"  # JSON serialization with compression
)
def get_user_profile(tenantId: str, user_id: str) -> dict:
    # Fetch user data from database
    return {"user_id": user_id, "name": "John Doe", "email": "john@example.com"}


# Dynamic TTL and cache bypass
@redis_memoize(
    memoize_key="dynamic_data",
    ttl=lambda tenantId, priority: 3600 if priority == "high" else 300,
    bypass_cache=lambda tenantId, force_refresh: force_refresh
)
def get_dynamic_data(tenantId: str, priority: str, force_refresh: bool = False) -> dict:
    return {"data": "fresh_data", "priority": priority}

Async Support

import asyncio
from rediskit import redis_memoize, init_async_redis_connection_pool

# Initialize async Redis connection pool
await init_async_redis_connection_pool()


@redis_memoize(memoize_key="async_calc", ttl=300)
async def async_expensive_function(tenantId: str, value: int) -> dict:
    await asyncio.sleep(1)  # Simulate async work
    return {"async_result": value * 100}


# Usage
result = await async_expensive_function("tenant1", 5)

Distributed Locking

from rediskit import get_redis_mutex_lock, get_async_redis_mutex_lock

# Synchronous distributed lock
with get_redis_mutex_lock("critical_section", expire=30) as lock:
    # Only one process can execute this block at a time
    perform_critical_operation()

# Async distributed lock
async with get_async_redis_mutex_lock("async_critical_section", expire=30) as lock:
    await perform_async_critical_operation()

Encryption Management

from rediskit import Encrypter

# Generate new encryption keys
encrypter = Encrypter()
new_key = encrypter.generate_new_hex_key()

# Encrypt/decrypt data manually
encrypted = encrypter.encrypt("sensitive data", useZstd=True)
decrypted = encrypter.decrypt(encrypted)

Configuration

Configure rediskit using environment variables:

# Redis connection settings
export REDISKIT_REDIS_HOST="localhost"
export REDISKIT_REDIS_PORT="6379"
export REDISKIT_REDIS_PASSWORD=""

# Encryption keys (base64-encoded JSON)
export REDISKIT_ENCRYPTION_SECRET="eyJfX2VuY192MSI6ICI0MGViODJlNWJhNTJiNmQ4..."

# Cache settings
export REDISKIT_REDIS_TOP_NODE="my_app_cache"
export REDISKIT_REDIS_SKIP_CACHING="false"

API Reference

Core Decorators

@RedisMemoize

Cache function results in Redis with configurable options.

Parameters:

  • memoizeKey: Cache key (string or callable)
  • ttl: Time to live in seconds (int, callable, or None)
  • bypassCache: Skip cache lookup (bool or callable)
  • cacheType: Serialization method ("zipJson" or "zipPickled")
  • resetTtlUponRead: Refresh TTL when reading from cache
  • enableEncryption: Encrypt cached data
  • storageType: Redis storage pattern ("string" or "hash")
  • connection: Custom Redis connection (optional)

Connection Management

  • init_redis_connection_pool(): Initialize sync Redis connection pool
  • init_async_redis_connection_pool(): Initialize async Redis connection pool
  • get_redis_connection(): Get sync Redis connection
  • get_async_redis_connection(): Get async Redis connection

Distributed Locking

  • GetRedisMutexLock(name, expire, auto_renewal, id): Get sync distributed lock
  • GetAsyncRedisMutexLock(name, expire, auto_renewal): Get async distributed lock

Encryption

  • Encrypter(keyHexDict): Encryption/decryption with key versioning

Requirements

  • Python 3.12+
  • Redis server
  • Dependencies: redis, redis-lock, nacl, zstd

License

Apache-2.0 license

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

rediskit-0.0.40.tar.gz (64.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

rediskit-0.0.40-py3-none-any.whl (45.5 kB view details)

Uploaded Python 3

File details

Details for the file rediskit-0.0.40.tar.gz.

File metadata

  • Download URL: rediskit-0.0.40.tar.gz
  • Upload date:
  • Size: 64.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for rediskit-0.0.40.tar.gz
Algorithm Hash digest
SHA256 96d8899d938ac68a3475841d360e69653d833ad3157936723720f0b38ecdd4be
MD5 59c79e008296cf9463c49d0560f077cf
BLAKE2b-256 f08cdf3ea59e95c986a62ecac962bf6dbe6322c0ab4bfa316a8db5cda68725e9

See more details on using hashes here.

Provenance

The following attestation bundles were made for rediskit-0.0.40.tar.gz:

Publisher: publish.yml on BadrElfarri/rediskit

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file rediskit-0.0.40-py3-none-any.whl.

File metadata

  • Download URL: rediskit-0.0.40-py3-none-any.whl
  • Upload date:
  • Size: 45.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for rediskit-0.0.40-py3-none-any.whl
Algorithm Hash digest
SHA256 6c6be800f68ee694201c843876465d29b17443f2cf13c52e0b4573799d7f59e9
MD5 4d2abfdf80bd8d17e70f2f90cccbad3c
BLAKE2b-256 1843c1106c75c1f5ba6f58594fc69f210ded9c3a897278d82a3096c0d33ebbbe

See more details on using hashes here.

Provenance

The following attestation bundles were made for rediskit-0.0.40-py3-none-any.whl:

Publisher: publish.yml on BadrElfarri/rediskit

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page