Skip to main content

A flexible task dispatcher for Python with multiple threading or processing control

Project description

Python Worker Dispatcher


A flexible task dispatcher for Python with multiple threading or processing control

PyPI

Features

  • Tasks Dispatching to managed workers

  • Elegant Interface for setup and use

  • Various modes to choose from


OUTLINE


DEMONSTRATION

Just write your own callback functions using the library, then run it and collect the result details:

$ python3 main.py

Worker Dispatcher Configutation:
- Local CPU core: 10
- Tasks Count: 100
- Runtime: Unlimited
- Dispatch Mode: Fixed Workers (Default)
- Workers Info:
   Worker Type: Processing
   Number of Workers : 10
   Max Worker: 10

--- Start to dispatch workers at 2024-06-14T17:46:30.996685+08:00 ---

...(User-defined output)...

--- End of worker dispatch at 2024-06-14T17:46:41.420888+08:00---

Spend Time: 10.424203 sec
Completed Tasks Count: 100
Uncompleted Tasks Count: 0
Undispatched Tasks Count: 0

Use 20 theads concurrently to dispatch tasks for HTTP reqeusts

import worker_dispatcher
import requests

def each_task(id: int, config, task, log):
    response = requests.get(config['my_endpoint'] + task)
    return response

responses = worker_dispatcher.start({
    'task': {
        'list': ['ORD_AH001', 'ORD_KL502', '...' , 'ORD_GR393'],
        'callback': each_task,
        'config': {
            'my_endpoint': 'https://your.name/order-handler/'
        },
    },
    'worker': {
        'number': 20,
    }
})

Utilizes all CPU cores on the machine to compute tasks.

import worker_dispatcher

def each_task(id: int, config, task, log):
    result = sum(id * i for i in range(10**9))
    return result

if __name__ == '__main__':
    results = worker_dispatcher.start({
        'task': {
            'list': 10,
            'callback': each_task,
        },
        'worker': {   
            'use_processing': True
        }
    })

INTRODUCTION

This library helps to efficiently consume tasks by using multiple threading or processing and returns all results jointly.

Introduction


INSTALLATION

To install the current release:

$ pip install worker-dispatcher

USAGE

By calling the start() method with the configuration parameters, the package will begin dispatching tasks while managing threading or processing based on the provided settings. Once the tasks are completed, the package will return all the results.

An example configuration setting with all options is as follows:

import worker_dispatcher 

results = worker_dispatcher.start({
    'debug': False,
    'task': {
        'list': [],                     # Support list and integer. Integer represent the number of tasks to be generated.
        'callback': callback_sample,
        'config': {},
        'result_callback': False
    },
    'worker': {
        'number': 8,
        'frequency_mode': {             # Changing from assigning tasks to a fixed number of workers once, to assigning tasks and workers frequently.
            'enabled': False, 
            'interval': 1,              # The second(s) of interval
            'accumulated_workers': 0,   # Accumulate the number of workers for each interval for next dispatch.
            'max_workers': None,        # limit the maximum number of workers to prevent system exhaustion.
        },
        'use_processing': False,        # To break GIL, workers will be based on processing pool.
        'parallel_processing': {        # To break GIL and require a number of workers greater than the number of CPU cores.
            'enabled': False,           # `worker.use_processing` setting will be ignored when enabled. The actual number of workers will be adjusted to a multiple of the CPU core count.
            'use_queue': False,         # Enable a task queue to specify the number of workers without adjustment, though the maximum may be limited by your device.
        },   
    },
    'runtime': None,                    # Dispatcher max runtime in seconds
    'verbose': True
})

Options

Option Type Deafult Description
debug bool False Debug mode
task.list multitype list The tasks for dispatching to each worker. *
- List: Each value will be passed as a parameter to your callback function.
- Integer: The number of tasks to be generated.
task.callback callable (sample) The callback function called by each worker runs
task.config multitype list The custom variable to be passed to the callback function
task.result_callback callable Null The callback function called when each task processes the result
worker.number int (auto) The number of workers to fork.
(The default value is the number of local CPU cores)
worker.frequency_mode.enabled bool False Changing from assigning tasks to a fixed number of workers once, to assigning tasks and workers frequently.
worker.frequency_mode.interval float 1 The second(s) of interval.
worker.frequency_mode.accumulated_workers int 0 Accumulate the number of workers for each interval for next dispatch.
worker.frequency_mode.max_workers int None limit the maximum number of workers to prevent system exhaustion.
worker.use_processing boolean False To break GIL, workers will be based on processing pool.
worker.parallel_processing.enabled bool False worker.use_processing setting will be ignored when enabled. The actual number of workers will be adjusted to a multiple of the CPU core count.
worker.parallel_processing.use_queue bool False Enable the use of a task queue instead of task dispatch, which allows specifying the number of workers but may be limited by your device.
runtime float None Dispatcher max runtime in seconds.
verbose bool True Enables or disables verbose mode for detailed output.

task.callback

The callback function called by each worker runs

callback_function (id: int, config, task, log: dict)
Argument Type Deafult Description
id int (auto) The sequence number generated by each task starting from 1
config multitype {} The custom variable to be passed to the callback function
task multitype (custom) Each value from the task.list
log dict {} The log from each task written by this callback function.

task.result_callback

The callback function called when each task processes the result

result_callback_function (id: int, config, result, log: dict)
Argument Type Deafult Description
id int (auto) The sequence number generated by each task starting from 1
config multitype {} The custom variable to be passed to the callback function
result multitype (custom) Each value returned back from task.callback
log dict (auto) Reference: get_logs()

Other Methods

  • get_results()

    Get all results in list type after completing start()

  • get_logs()

    Get all logs in list type after completing start()

    Each log is of type dict, containing the results of every task processed by the worker:

    • task_id
    • started_at
    • ended_at
    • duration
    • result
  • get_result_info()

    Get a dict with the whole spending time and started/ended timestamps after completing start()

  • get_tps()

    Get TPS report in dict type after completing start() or by passing a list data.

    def get_tps(logs: dict=None, debug: bool=False, interval: float=0, reverse_interval: bool = False, display_intervals: bool = False) -> dict:
    

    The log dict matches the format of the get_logs() and refers to it by default. Each result of a log will be considered valid if it meets one of the following conditions:

    • It is a requests.Response object with a status code of 200
    • It is a valid value other than the aforementioned object

Scenarios

Stress Test

Perform a stress test scenario with 10 requests per second.

import worker_dispatcher, requests

def each_task(id, config, task, log):
    response = None
    try:
        response = requests.get(config['my_endpoint'], timeout=(5, 10))
    except requests.exceptions.RequestException as e:
        print("An error occurred:", e)
    return response

responses = worker_dispatcher.start({
    'task': {
        'list': 5000,
        'callback': each_task,
        'config': {
            'my_endpoint': 'https://your.name/api'
        },
    },
    'worker': {
        'number': 10,
        'frequency_mode': {
            'enabled': True, 
            'interval': 1,
        },
    }
})

print(worker_dispatcher.get_logs())
print(worker_dispatcher.get_tps())

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

worker_dispatcher-1.0.5.tar.gz (14.1 kB view details)

Uploaded Source

Built Distribution

worker_dispatcher-1.0.5-py3-none-any.whl (12.8 kB view details)

Uploaded Python 3

File details

Details for the file worker_dispatcher-1.0.5.tar.gz.

File metadata

  • Download URL: worker_dispatcher-1.0.5.tar.gz
  • Upload date:
  • Size: 14.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.9.20

File hashes

Hashes for worker_dispatcher-1.0.5.tar.gz
Algorithm Hash digest
SHA256 fc1eb1afe81b44dde7d59a420c42d2f6a4676e339f4a7b43c3f8d7862177401d
MD5 48260a2e60aa876470669f96e68bebd6
BLAKE2b-256 abfad1365f2c58b73d3ad7b1071dc8c8eb63617b0a06729796f2e283f210a9d5

See more details on using hashes here.

File details

Details for the file worker_dispatcher-1.0.5-py3-none-any.whl.

File metadata

File hashes

Hashes for worker_dispatcher-1.0.5-py3-none-any.whl
Algorithm Hash digest
SHA256 6c1f2b52ce9d72cc3028bdc189c24284d121668c79f5c5712eb0ca6a54f8348e
MD5 a92849e30b6aeca72d3ceb4f39c7ec78
BLAKE2b-256 c33143f272e2931da67667088ed3e14e0f54e31f134c618aed4165d298889bf8

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page