Skip to main content

Dynamically created, cached mutexes in Python.

Project description


A small utility library for dynamically creating mutexes based on cache keys.

Use case

Say you're implementing a graphql server, with object fields which asynchronously resolve independently of each other. Multiple fields perform the same operation, and thus should use a mutex and some basic caching to ensure that the database query happens at most once. If you have an array of these objects, suddenly they're all using the same mutex, which can degrade performance.

Instead of using one mutex to rule them all, dynamically create multiple short-lived mutexes which each object can use independently of other resolving objects. By using the same cache key for the mutexes as you would for your cache check, you can almost transparently use dynamically created mutexes without worrying about performance or allocation/deallocation of mutexes.

With mutexcache, if a mutex associated with a cache key is already stored, then it will be returned. Otherwise, a new mutex will silently be created, stored for future use, and returned.


pip3 install mutexcache


MutexCache.get() returns threading.Lock objects.

from mutexcache import MutexCache

mutex_cache = MutexCache() # optionally, provide a ttl in seconds 

cache_key_a = "key_a"
cache_key_b = "key_b"

mut_a = mutex_cache.get(cache_key_a)
mut_b = mutex_cache.get(cache_key_b)

# [ stuff with your Locks...]

Project details

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mutexcache-0.1.1.tar.gz (2.3 kB view hashes)

Uploaded Source

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page