Skip to main content

No project description provided

Project description

kube_watch

To setup the project:

  • poetry install
  • poetry shell

To install package to your environment locally:

python setup.py install

To publish

poetry config pypi-token.pypi your-api-token poetry build poetry publish

Description

The kube_watch library is build on top of Prefect. The library is designed to define workflows in a declaritive and flexible fashion. Originally, workflows in Prefect are defined via decorators such as @flow and @task. In kube_watch, workflows can be defined in a declaritive form via yaml files. The library is mainly focused on running scheduled workflows in kubernetes environment. However, it can easily be extended to be used for any purpose requiring a workflow. The workflow manifest has the following generic structure:

workflow:
  name: Dummy Workflow
  runner: concurrent
  tasks:
    - name: Task_A
      module: <module_path>
      task: <func_name>
      inputsArgType: arg
      inputs:
        parameters:
          - name: x1
            value: y1
          - name: x2
            value: y2
          - name: x3
            type: env
            value: Y3

    - name: Task_B
      module: <module_path>
      task: <func_name>
      inputsArgType: arg
      inputs:
        parameters:
          - name: xx1
            value: yy1
          - name: xx2
            value: yy2
      dependency:
        - taskName: Task_A
          inputParamName: xx3

    - name: Task_C
      module: <module_path>
      task: <func_name>
      inputsArgType: arg
      conditional:
        tasks: ["Task_B"]

    - name: Task_external
      module: <module_external_path>
      plugin_path: <path_to_module>
      task: <func_name>
      inputsArgType: arg
      conditional:
        tasks: ["Task_B"]

runner: concurrent | sequential: if concurrent selected, tasks will be run concurrently.

module: all modules are located in 'modules' directory in kube_watch. This is where you can extend the library and add new tasks / modules. Below modules, there are submodules such as providers, clusters, and logic. Within each of this submodules, specific modules are defined. For example: providers.aws contains a series of tasks related to AWS. In this case, <module_path> = providers.aws. To add new tasks, add a new module with a similar pattern and refer the path in your task block.

plugin_path: you can add new funtions outside the library without needing to modify this library. If plugin_path is defined, the library assumes you are referring to a module outside the library (e.g. in your app or workflow). The plugin_path is path to your py script (the folder). Then, provide your script name as module and function name as task.

task: task is simply the name function that you put in the <module_path>. i.e. as you define a function in a module, you can simply start to use it in your manifests.

inputArgType: arg | dict | list: if the task functions accept known-fixed number of parameters, then use arg.

dependency: this block defines dependency of a child task to its parent. If inputParamName is defined, then OUTPUT of the parent task is passed to the child with an argument name defined by inputParamName.

IMPORTATN NOTE: A strict assumption is that task functions return a single output. If there are cases with multiple output, wrap them into a dictionary and unwrap them in the child task.

conditional: These are blocks where you can define when a task runs depending on the outcome of its parent. The parent task should return True or False.

Parameters have also a type entry: env | static. static is default value. If type is defined as env, the parameter value is loaded from Environment Variables. In this case, value should be the name of the corresponding env var.

In above examples:

def Task_A(x1, x2, x3):
    # do something
    return output_A

def Task_B(xx1, xx2, xx3):
    # do something else
    return output_B

def Task_C():
    # do another thing
    return output_C

Batch workflows

kube_watch also enables to run workflows in batch. A separate manifest with following form is required:

batchFlows: runner: sequential items: - path: path_to_flow_A.yaml - path: path_to_flow_B.yaml - path: path_to_flow_C.yaml

cron_app

The cron_app folder contains an example use case of kube_watch library. The cron_app can be used to deploy a CronJob in a kubernetes environment. The app assumes the manifests are located in a separate repository. It will clone the repo and read the manifests and runs the workflows.

Connect to a server

Start Server

prefect server start

To Connect

To connect to a server, simply set the following environment variable: PREFECT_API_URL

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

kubernetes_watch-0.1.15.tar.gz (29.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

kubernetes_watch-0.1.15-py3-none-any.whl (36.0 kB view details)

Uploaded Python 3

File details

Details for the file kubernetes_watch-0.1.15.tar.gz.

File metadata

  • Download URL: kubernetes_watch-0.1.15.tar.gz
  • Upload date:
  • Size: 29.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for kubernetes_watch-0.1.15.tar.gz
Algorithm Hash digest
SHA256 cf3a94f3b417ec7105c8abfa06cdb712cd2fbad3e6b4579a67c7af1950bf3043
MD5 3a1bd08ed26f0a951be3deb0740a04c9
BLAKE2b-256 fc1d45ef86467a19bb7a93c365976dc111349992176477586263fcdc8bd0be79

See more details on using hashes here.

File details

Details for the file kubernetes_watch-0.1.15-py3-none-any.whl.

File metadata

File hashes

Hashes for kubernetes_watch-0.1.15-py3-none-any.whl
Algorithm Hash digest
SHA256 4069e4d1a06f55f031f0458991cd8d811792475e381d290e27837b63e17f1223
MD5 f6dec1ed3f2607ccb47981d4cd71c0ef
BLAKE2b-256 9d43f54257e695018aac63aa182c09f40448ca826cce95a7e1dd6b7ccf8d4e81

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page