Skip to main content

No project description provided

Project description

kube_watch

To setup the project:

  • poetry install
  • poetry shell

To install package to your environment locally:

python setup.py install

To publish

poetry config pypi-token.pypi your-api-token poetry build poetry publish

Description

The kube_watch library is build on top of Prefect. The library is designed to define workflows in a declaritive and flexible fashion. Originally, workflows in Prefect are defined via decorators such as @flow and @task. In kube_watch, workflows can be defined in a declaritive form via yaml files. The library is mainly focused on running scheduled workflows in kubernetes environment. However, it can easily be extended to be used for any purpose requiring a workflow. The workflow manifest has the following generic structure:

workflow:
  name: Dummy Workflow
  runner: concurrent
  tasks:
    - name: Task_A
      module: <module_path>
      task: <func_name>
      inputsArgType: arg
      inputs:
        parameters:
          - name: x1
            value: y1
          - name: x2
            value: y2
          - name: x3
            type: env
            value: Y3

    - name: Task_B
      module: <module_path>
      task: <func_name>
      inputsArgType: arg
      inputs:
        parameters:
          - name: xx1
            value: yy1
          - name: xx2
            value: yy2
      dependency:
        - taskName: Task_A
          inputParamName: xx3

    - name: Task_C
      module: <module_path>
      task: <func_name>
      inputsArgType: arg
      conditional:
        tasks: ["Task_B"]

    - name: Task_external
      module: <module_external_path>
      plugin_path: <path_to_module>
      task: <func_name>
      inputsArgType: arg
      conditional:
        tasks: ["Task_B"]

runner: concurrent | sequential: if concurrent selected, tasks will be run concurrently.

module: all modules are located in 'modules' directory in kube_watch. This is where you can extend the library and add new tasks / modules. Below modules, there are submodules such as providers, clusters, and logic. Within each of this submodules, specific modules are defined. For example: providers.aws contains a series of tasks related to AWS. In this case, <module_path> = providers.aws. To add new tasks, add a new module with a similar pattern and refer the path in your task block.

plugin_path: you can add new funtions outside the library without needing to modify this library. If plugin_path is defined, the library assumes you are referring to a module outside the library (e.g. in your app or workflow). The plugin_path is path to your py script (the folder). Then, provide your script name as module and function name as task.

task: task is simply the name function that you put in the <module_path>. i.e. as you define a function in a module, you can simply start to use it in your manifests.

inputArgType: arg | dict | list: if the task functions accept known-fixed number of parameters, then use arg.

dependency: this block defines dependency of a child task to its parent. If inputParamName is defined, then OUTPUT of the parent task is passed to the child with an argument name defined by inputParamName.

IMPORTATN NOTE: A strict assumption is that task functions return a single output. If there are cases with multiple output, wrap them into a dictionary and unwrap them in the child task.

conditional: These are blocks where you can define when a task runs depending on the outcome of its parent. The parent task should return True or False.

Parameters have also a type entry: env | static. static is default value. If type is defined as env, the parameter value is loaded from Environment Variables. In this case, value should be the name of the corresponding env var.

In above examples:

def Task_A(x1, x2, x3):
    # do something
    return output_A

def Task_B(xx1, xx2, xx3):
    # do something else
    return output_B

def Task_C():
    # do another thing
    return output_C

Batch workflows

kube_watch also enables to run workflows in batch. A separate manifest with following form is required:

batchFlows: runner: sequential items: - path: path_to_flow_A.yaml - path: path_to_flow_B.yaml - path: path_to_flow_C.yaml

cron_app

The cron_app folder contains an example use case of kube_watch library. The cron_app can be used to deploy a CronJob in a kubernetes environment. The app assumes the manifests are located in a separate repository. It will clone the repo and read the manifests and runs the workflows.

Connect to a server

Start Server

prefect server start

To Connect

To connect to a server, simply set the following environment variable: PREFECT_API_URL

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

kubernetes_watch-0.1.11.tar.gz (27.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

kubernetes_watch-0.1.11-py3-none-any.whl (33.5 kB view details)

Uploaded Python 3

File details

Details for the file kubernetes_watch-0.1.11.tar.gz.

File metadata

  • Download URL: kubernetes_watch-0.1.11.tar.gz
  • Upload date:
  • Size: 27.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.11

File hashes

Hashes for kubernetes_watch-0.1.11.tar.gz
Algorithm Hash digest
SHA256 ae22ddad032b8bd395f57ce4ea80894465d0b0f9d86536301919fa1e5f5da04e
MD5 c60683fe1435eb122f4ede9291bc1848
BLAKE2b-256 580eb314cc99dfec81b2c9567d2f8cf69ed8e1f965c3b584408990975c6226f2

See more details on using hashes here.

File details

Details for the file kubernetes_watch-0.1.11-py3-none-any.whl.

File metadata

File hashes

Hashes for kubernetes_watch-0.1.11-py3-none-any.whl
Algorithm Hash digest
SHA256 10721f8a31ca87a0b56c6621cd9ff54f5a5c7d0f4d3f0dd0c78e307a43323d24
MD5 30376be39573f33dfb4e2c877378cb59
BLAKE2b-256 9790a5e6fe56081751fe9158431c73a79079682cc789f74cd8f1cd9cce44ccdf

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page