Skip to main content

A small, powerful CLI coding agent for open AI models

Project description

Swival Logo

Swival

A coding agent for open models. Documentation

Swival connects to LM Studio, HuggingFace Inference API, or OpenRouter, sends your task, and runs an autonomous tool loop until it produces an answer. With LM Studio it auto-discovers your loaded model, so there's nothing to configure. A few thousand lines of Python, no framework.

Quickstart

LM Studio

  1. Install LM Studio and load a model with tool-calling support. Recommended first model: qwen3-coder-next (great quality/speed tradeoff on local hardware). Crank the context size as high as your hardware allows.
  2. Start the LM Studio server.
  3. Install Swival:
uv tool install swival
  1. Run:
swival "Refactor the error handling in src/api.py"

That's it. Swival finds the model, connects, and goes to work.

HuggingFace

export HF_TOKEN=hf_...
uv tool install swival
swival "Refactor the error handling in src/api.py" \
    --provider huggingface --model zai-org/GLM-5

You can also point it at a dedicated endpoint with --base-url and --api-key.

OpenRouter

export OPENROUTER_API_KEY=sk_or_...
uv tool install swival
swival "Refactor the error handling in src/api.py" \
    --provider openrouter --model openrouter/free

Interactive sessions

swival --repl

The REPL carries conversation history across questions, which makes it good for exploratory work and longer tasks.

Updates and uninstall

uv tool upgrade swival    # update
uv tool uninstall swival  # remove

What makes it different

Your models, your way. Swival works with LM Studio and HuggingFace Inference API. With LM Studio, it auto-discovers whatever model you have loaded. With HuggingFace, point it at any supported model or your own dedicated endpoint. You pick the model and the infrastructure.

Small enough to read and hack. The whole agent is a few thousand lines of Python across a handful of files, with no framework underneath. You can read the entire thing in an afternoon. If something doesn't work the way you want, you can change it.

Structured thinking for any model. The built-in think tool gives any model (including local ones) multi-step reasoning with revisions, branches, and persistent notes that survive context compaction. It's not locked to a specific provider.

Built for benchmarking. Pass --report report.json and Swival writes a machine-readable evaluation report with per-call LLM timing, tool success/failure counts, context compaction events, and guardrail interventions. Good for comparing models systematically on real coding tasks.

CLI-native. stdout is exclusively the final answer. All diagnostics go to stderr. You can pipe Swival's output straight into another command or a file.

Documentation

Full documentation is available at swival.github.io/swival.

  • Getting Started -- installation, first run, what happens under the hood
  • Usage -- one-shot mode, REPL mode, CLI flags, piping, exit codes
  • Tools -- what the agent can do: file ops, search, editing, web fetching, thinking, command execution
  • Safety and Sandboxing -- path resolution, symlink protection, command whitelisting, YOLO mode
  • Skills -- creating and using SKILL.md-based agent skills
  • Customization -- project instructions, system prompt overrides, tuning parameters
  • Providers -- LM Studio, HuggingFace, and OpenRouter configuration
  • Reports -- JSON reports for benchmarking and evaluation
  • Reviews -- external reviewer scripts for automated QA and LLM-as-judge evaluation
  • Using Swival with AgentFS -- copy-on-write filesystem sandboxing for safe agent runs

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

swival-0.1.4.tar.gz (687.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

swival-0.1.4-py3-none-any.whl (49.5 kB view details)

Uploaded Python 3

File details

Details for the file swival-0.1.4.tar.gz.

File metadata

  • Download URL: swival-0.1.4.tar.gz
  • Upload date:
  • Size: 687.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for swival-0.1.4.tar.gz
Algorithm Hash digest
SHA256 02f32b1f39cd988576a642970ae70c341304b444e8d1209e8480658722edb418
MD5 857e2a95a41d7d098c9abe1a9263f0e6
BLAKE2b-256 b77e6a0461478a68e613095a227b0ddf43d25ef75f793bee88e7795b7d84dddf

See more details on using hashes here.

File details

Details for the file swival-0.1.4-py3-none-any.whl.

File metadata

  • Download URL: swival-0.1.4-py3-none-any.whl
  • Upload date:
  • Size: 49.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for swival-0.1.4-py3-none-any.whl
Algorithm Hash digest
SHA256 2a5207320fa788692e6a533535635e8948e902a394a1f0df3b8f62e5215827dd
MD5 ce8751807e64701faeca635fe32afcd5
BLAKE2b-256 607872b22988fc98b1a040db7ce48c9a617be2cabe323b9335de6dad31b43d58

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page