Skip to main content

Scale Distribution Framework

Project description

Scaled

This project is aiming the target that provides simple and efficient and reliable way for distributing computing framework, centralized scheduler and stable protocol when client and worker talking to scheduler

Introduction

The goal for this project should be as simple as possible

  • It built on top of zmq
  • it has ready python version of Client, Scheduler, Worker
  • I will provide golang or Rust version of Scheduler, the goal for the Scheduler should be completely computer language agnostic, which means they follow the same protocol
  • Scheduler might support function based computing tree in the future

Installation

pip install scaled

if you want to use uvloop, please do: pip install uvloop, default we are using python builtin uvloop

How to use it

Start local scheduler and cluster at the same time in the code

import random

from scaled.client.client import Client
from scaled.cluster.combo import SchedulerClusterCombo
from scaled.utility.zmq_config import ZMQConfig, ZMQType
from scaled.utility.logging.scoped_logger import ScopedLogger
from scaled.utility.logging.utility import setup_logger


def calculate(sec: int):
  return sec * 1


def main():
  setup_logger()
  config = ZMQConfig(type=ZMQType.tcp, host="127.0.0.1", port=2345)

  cluster = SchedulerClusterCombo(address=config, n_workers=10, event_loop="uvloop")
  client = Client(config=config)

  tasks = [random.randint(0, 100) for _ in range(100000)]

  with ScopedLogger(f"submit {len(tasks)} tasks"):
    futures = [client.submit(calculate, i) for i in tasks]

  with ScopedLogger(f"gather {len(futures)} results"):
    results = [future.result() for future in futures]

  assert results == tasks

  cluster.shutdown()
  client.disconnect()


if __name__ == "__main__":
  main()

Start scheduler and cluster independently

use scaled_scheduler to start scheduler, for example:

scaled_scheduler --allocator-type queued tcp://0.0.0.0:8516

use scaled_cluster to start workers:

scaled_worker -n 10 tcp://127.0.0.1:8516

Then you can write simply write client code as:

from scaled.utility.zmq_config import ZMQConfig, ZMQType
from scaled.client.client import Client

def foobar(foo: int):
    return foo

config = ZMQConfig(type=ZMQType.tcp, host="127.0.0.1", port=2345)
client = Client(config=config)
future = client.submit(foobar, 1)

print(future.result())

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

scaled-0.5.tar.gz (23.2 kB view details)

Uploaded Source

Built Distribution

scaled-0.5-py3-none-any.whl (35.1 kB view details)

Uploaded Python 3

File details

Details for the file scaled-0.5.tar.gz.

File metadata

  • Download URL: scaled-0.5.tar.gz
  • Upload date:
  • Size: 23.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.1

File hashes

Hashes for scaled-0.5.tar.gz
Algorithm Hash digest
SHA256 8d6162b58696b8bee8e7888438d89d88e922f1da875373aed629c7a53dd0f405
MD5 087b51cdbaf386d91865beb7bcb8f5b9
BLAKE2b-256 566d0fe069567372765db967305aa91fffbf368a3c07c9428d09e370740d6804

See more details on using hashes here.

File details

Details for the file scaled-0.5-py3-none-any.whl.

File metadata

  • Download URL: scaled-0.5-py3-none-any.whl
  • Upload date:
  • Size: 35.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.1

File hashes

Hashes for scaled-0.5-py3-none-any.whl
Algorithm Hash digest
SHA256 14711687b853911a3aabb916c3bc8821ef6f46ccde4fcbfe91057775d853c549
MD5 465fa7317ac9b2340a5dc93e3fbc503b
BLAKE2b-256 304fc1549956bbf6d2c1ee85e7ac8cd18645e527d02541585b46ecf873006ec6

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page