Async tools for Python
Async Tools for Python.
Table of Contents
Threading is the most simple thing, but because of GIL it’s useless for computation. Only use when you want to parallelize the access to a blocking resource, e.g. network.
Decorator for functions that should be run in a separate thread. When the function is called, it returns a `threading.Event <https://docs.python.org/2/library/threading.html#event-objects>`__.
from asynctools.threading import Async @Async def request(url): # ... do request request('http://example.com') # Async request request('http://example.com').wait() # wait for it to complete
If you want to wait for multiple threads to complete, see next chapters.
Execute functions in parallel and collect results. Each function is executed in its own thread, all threads exit immediately.
- __call__(*args, **kwargs): Add a job. Call the Parallel object so it calls the worker function with the same arguments
- map(jobs): Convenience method to call the worker for every argument
- first(timeout=None): Wait for a single result to be available, with an optional timeout in seconds. The result is returned as soon as it’s ready. If all threads fail with an error – None is returned.
- join(): Wait for all tasks to be finished, and return two lists:
- A list of results
- A list of exceptions
from asynctools.threading import Parallel def request(url): # ... do request return data # Execute pll = Parallel(request) for url in links: pll(url) # Starts a new thread # Wait for the results results, errors = pll.join()
Since the request method takes just one argument, this can be chained:
results, errors = Parallel(request).map(links).join()
Create a pool of threads and execute work in it. Useful if you do want to launch a limited number of long-living threads.
Methods are same with `Parallel <#parallel>`__, with some additions:
- __call__(*args, **kwargs)
- close(): Terminate all threads. The pool is no more usable when closed.
- __enter__, __exit__ context manager to be used with with statement
from asynctools.threading import Pool def request(url): # ... do long request return data # Make pool pool = Pool(request, 5) # Assign some job for url in links: pll(url) # Runs in a pool # Wait for the results results, errors = pll.join()
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
|Filename, Size & Hash SHA256 Hash Help||File Type||Python Version||Upload Date|
(7.9 kB) Copy SHA256 Hash SHA256
|Wheel||2.7||Aug 14, 2014|
(6.3 kB) Copy SHA256 Hash SHA256
|Source||None||Aug 14, 2014|