Skip to main content

Scale Distribution Framework

Project description

Scaled

This project is aiming the target that provides simple and efficient and reliable way for distributing computing framework, centralized scheduler and stable protocol when client and worker talking to scheduler

Introduction

The goal for this project should be as simple as possible

  • It built on top of zmq
  • it has ready python version of Client, Scheduler, Worker
  • I will provide golang or Rust version of Scheduler, the goal for the Scheduler should be completely computer language agnostic, which means they follow the same protocol
  • Scheduler might support function based computing tree in the future

Installation

pip install scaled

if you want to use uvloop, please do: pip install uvloop, default we are using python builtin uvloop

How to use it

Start local scheduler and cluster at the same time in the code

import random

from scaled.client.client import Client
from scaled.cluster.combo import SchedulerClusterCombo
from scaled.utility.zmq_config import ZMQConfig, ZMQType
from scaled.utility.logging.scoped_logger import ScopedLogger
from scaled.utility.logging.utility import setup_logger


def calculate(sec: int):
  return sec * 1


def main():
  setup_logger()
  config = ZMQConfig(type=ZMQType.tcp, host="127.0.0.1", port=2345)

  cluster = SchedulerClusterCombo(address=config, n_workers=10, event_loop="uvloop")
  client = Client(config=config)

  tasks = [random.randint(0, 100) for _ in range(100000)]

  with ScopedLogger(f"submit {len(tasks)} tasks"):
    futures = [client.submit(calculate, i) for i in tasks]

  with ScopedLogger(f"gather {len(futures)} results"):
    results = [future.result() for future in futures]

  assert results == tasks

  cluster.shutdown()
  client.disconnect()


if __name__ == "__main__":
  main()

Start scheduler and cluster independently

use scaled_scheduler to start scheduler, for example:

scaled_scheduler --allocator-type queued tcp://0.0.0.0:8516

use scaled_cluster to start workers:

scaled_worker -n 10 tcp://127.0.0.1:8516

Then you can write simply write client code as:

from scaled.utility.zmq_config import ZMQConfig, ZMQType
from scaled.client.client import Client

def foobar(foo: int):
    return foo

config = ZMQConfig(type=ZMQType.tcp, host="127.0.0.1", port=2345)
client = Client(config=config)
future = client.submit(foobar, 1)

print(future.result())

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

scaled-0.7.tar.gz (23.3 kB view details)

Uploaded Source

Built Distribution

scaled-0.7-py3-none-any.whl (35.3 kB view details)

Uploaded Python 3

File details

Details for the file scaled-0.7.tar.gz.

File metadata

  • Download URL: scaled-0.7.tar.gz
  • Upload date:
  • Size: 23.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.1

File hashes

Hashes for scaled-0.7.tar.gz
Algorithm Hash digest
SHA256 611f84d48e296ac59b726d3946b503931f7e882279f1f7d89e6283cc51fa11ea
MD5 f5c24a8a152c07182e490f23f2bb438c
BLAKE2b-256 d05e66c19fded924369c6cf0a7e941b55a0cab4b21966ced8c15c6fdc3ca97e5

See more details on using hashes here.

File details

Details for the file scaled-0.7-py3-none-any.whl.

File metadata

  • Download URL: scaled-0.7-py3-none-any.whl
  • Upload date:
  • Size: 35.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.1

File hashes

Hashes for scaled-0.7-py3-none-any.whl
Algorithm Hash digest
SHA256 feb20c406a1eb35d5ac6c6224ad871178191eaa20fc18955f60ec3b8d65e4f69
MD5 74384ad4daf866bcf6e4739095dcdfb7
BLAKE2b-256 a6f457ad0560a360ac3b22139e827a730b55be435b634886b68af841f33ae8fd

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page