Skip to main content

Run Harbor tasks in the cloud with scheduling, monitoring, and persistent state

Project description

Oddish CLI

Run Harbor tasks on Oddish infrastructure.

oddish is a Python CLI for submitting Harbor tasks, running multi-trial sweeps, monitoring experiments, and pulling logs and artifacts back to disk. If you already use harbor run, Oddish adds persistent state, retries, queueing, and better operational tooling around the same task format.

Python 3.12+ is required.

Quick Start

uv pip install oddish

export ODDISH_API_KEY="ok_..."

# Submit a run
oddish run -d swebench@1.0 -a codex -m openai/gpt-5.2 --n-trials 3

# Watch progress
oddish status
oddish status <task_id> --watch

# Pull logs and artifacts locally
oddish pull <task_id> --watch

The CLI targets Oddish Cloud by default. All API-backed commands require ODDISH_API_KEY. For self-deployed instances, also set ODDISH_API_URL.

Installation

uv pip install oddish

Common environment variables:

export ODDISH_API_KEY="ok_..."

# Point at a self-deployed instance instead of Oddish Cloud
# export ODDISH_API_URL="https://<workspace>--api.modal.run"

# Optional dashboard override
# export ODDISH_DASHBOARD_URL="https://www.oddish.app"

Need to deploy your own stack? See ../SELF_HOSTING.md. Need package internals, architecture, or development notes? See AGENTS.md.

Commands

The installed console script is:

oddish --help

Available commands:

  • oddish run uploads a local task or dataset, downloads a registry dataset, or expands a sweep config into trials
  • oddish status shows system, task, or experiment status
  • oddish cancel stops all in-flight runs for a task
  • oddish pull downloads logs, results, trajectories, and artifact files for a trial, task, or experiment
  • oddish delete deletes a task or experiment from a self-hosted deployment

oddish run

Use oddish run for:

  • a single local Harbor task directory
  • a local dataset directory containing multiple tasks
  • a Harbor registry dataset via --dataset
  • a YAML or JSON sweep config via --config
  • appending trials to an existing task via --task

Examples:

# Local task
oddish run ./my-task -a claude-code -m anthropic/claude-sonnet-4-5

# Local dataset
oddish run ./my-dataset -a codex -m openai/gpt-5.2 --n-trials 3

# Harbor registry dataset
oddish run -d swebench@1.0 -a codex -m openai/gpt-5.2 --n-trials 3

# Filter a dataset
oddish run -d swebench@1.0 -t "django__*" -l 10 -a claude-code

# Append new trials to an existing task
oddish run --task task_123 -a gemini-cli -m google/gemini-3.1-pro-preview --n-trials 3

# Submit in the background
oddish run ./my-task -a claude-code --background

# Script-friendly JSON output (implies --background)
oddish run ./my-task -a claude-code --json

Common flags:

  • PATH or -p, --path selects a local task or dataset directory
  • -a, --agent selects the agent
  • -m, --model selects the model
  • --n-trials runs multiple trials per task
  • -d, --dataset pulls tasks from the Harbor registry
  • --task appends trials to an existing task ID without re-uploading task files
  • -c, --config loads a YAML or JSON sweep config
  • -t, --task-name, -x, --exclude-task-name, and -l, --n-tasks filter datasets
  • -e, --env selects the execution environment
  • -P, --priority, -E, --experiment, -u, --user, -G, --github-user, and --github-meta attach scheduling and attribution metadata
  • -w, --watch / --no-watch watches single-task submissions until completion
  • --background submits and returns immediately
  • --json emits machine-readable output and implies --background
  • -q, --quiet suppresses nonessential output
  • --run-analysis runs post-trial analysis and task verdict generation
  • --publish publishes the experiment for public read-only access
  • --disable-verification skips task verification
  • --override-cpus, --override-memory-mb, --override-gpus, --override-storage-mb, and --force-build override environment settings
  • --ae/--agent-env, --ak/--agent-kwarg, and --artifact pass Harbor agent/env configuration through to every submitted config
  • --api overrides the API URL for a single invocation

Supported --env values:

  • docker
  • daytona
  • e2b
  • modal
  • runloop
  • gke

When --env is omitted:

  • hosted Oddish (*.modal.run) defaults to modal
  • other API URLs default to docker
  • --task preserves the existing task's environment unless you override it

Sweep Configs

oddish run -c sweep.yaml accepts YAML or JSON. A minimal config:

agents:
  - name: claude-code
    model_name: anthropic/claude-sonnet-4-5
    n_trials: 3
  - name: codex
    model_name: openai/gpt-5.2
    n_trials: 3

dataset: swebench@1.0
n_tasks: 10
priority: low

You can also set path, exclude_task_names, and experiment_id in the config file. Per-agent overrides use env and kwargs. Timeouts and per-provider concurrency are no longer configured in sweep files; declare task timeouts in task.toml and API concurrency at server startup.

oddish status

Without arguments, oddish status shows recent experiments and API health. Use a task ID or --experiment to inspect a specific run, and --watch to resume live monitoring later.

Examples:

# System overview
oddish status

# Task snapshot
oddish status <task_id>

# Watch a task
oddish status <task_id> --watch

# Watch an experiment
oddish status --experiment <experiment_id> --watch

oddish cancel

Cancel all in-flight runs for a task without deleting any data. Queued jobs are removed, running trials are cancelled, and active Modal workers are terminated when applicable. Completed trials and their results are preserved.

oddish cancel <task_id>
oddish cancel <task_id> --force   # skip confirmation

oddish pull

oddish pull accepts a trial ID, task ID, or experiment ID and auto-detects the target type by default.

Examples:

# Pull one trial
oddish pull <trial_id>

# Keep syncing a task while it runs
oddish pull <task_id> --watch --interval 5

# Pull an entire experiment, including task files
oddish pull <experiment_id> --include-task-files

By default, pull output is written to ./.oddish/<target> and includes a manifest.json describing the fetch. Use --no-logs, --no-files, --structured, --include-task-files, --out, and --type to control what gets downloaded and where it lands. --type trial|task|experiment forces the target type instead of auto-resolving it.

oddish delete

Examples:

# Delete a task and its trials
oddish delete <task_id>

# Delete an entire experiment
oddish delete --experiment <experiment_id>

Typical Workflow

# 1. Submit a run
oddish run -d swebench@1.0 -a claude-code -m anthropic/claude-sonnet-4-5

# 2. Inspect or watch it later
oddish status <task_id> --watch

# 3. Pull outputs when you want them locally
oddish pull <task_id> --watch

More Technical Docs

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

oddish-0.1.8.tar.gz (18.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

oddish-0.1.8-py3-none-any.whl (21.0 kB view details)

Uploaded Python 3

File details

Details for the file oddish-0.1.8.tar.gz.

File metadata

  • Download URL: oddish-0.1.8.tar.gz
  • Upload date:
  • Size: 18.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.9

File hashes

Hashes for oddish-0.1.8.tar.gz
Algorithm Hash digest
SHA256 3fb6f2932209a785c4c78a8cef3e5872eebf946e8e9b5f260b4c1cf032591dee
MD5 efa963364fe13b59695978cf232906f2
BLAKE2b-256 83a3e2558f776761bb6346c41488636f359ea7e552cf3f40db8ebd018f1ebfb3

See more details on using hashes here.

File details

Details for the file oddish-0.1.8-py3-none-any.whl.

File metadata

  • Download URL: oddish-0.1.8-py3-none-any.whl
  • Upload date:
  • Size: 21.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.9

File hashes

Hashes for oddish-0.1.8-py3-none-any.whl
Algorithm Hash digest
SHA256 2d7304bccabc08289c44514f31f84900255c0a0ab208431d1db86081bc5e65c9
MD5 c75240a0ddda240a44806eb6f1d86eff
BLAKE2b-256 4301a1e88e51052b539a9b20d61d19e4be1d72402bf50043b4f8f239c1908330

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page